userfaultfd: use per-vma locks in userfaultfd operations
All userfaultfd operations, except write-protect, opportunistically use per-vma locks to lock vmas. On failure, attempt again inside mmap_lock critical section. Write-protect operation requires mmap_lock as it iterates over multiple vmas. Link: https://lkml.kernel.org/r/20240215182756.3448972-5-lokeshgidra@google.com Signed-off-by: Lokesh Gidra <lokeshgidra@google.com> Reviewed-by: Liam R. Howlett <Liam.Howlett@oracle.com> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Axel Rasmussen <axelrasmussen@google.com> Cc: Brian Geffon <bgeffon@google.com> Cc: David Hildenbrand <david@redhat.com> Cc: Jann Horn <jannh@google.com> Cc: Kalesh Singh <kaleshsingh@google.com> Cc: Matthew Wilcox (Oracle) <willy@infradead.org> Cc: Mike Rapoport (IBM) <rppt@kernel.org> Cc: Nicolas Geoffray <ngeoffray@google.com> Cc: Peter Xu <peterx@redhat.com> Cc: Ryan Roberts <ryan.roberts@arm.com> Cc: Suren Baghdasaryan <surenb@google.com> Cc: Tim Murray <timmurray@google.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
32af81af2f
commit
867a43a34f
@ -2005,17 +2005,8 @@ static int userfaultfd_move(struct userfaultfd_ctx *ctx,
|
||||
return -EINVAL;
|
||||
|
||||
if (mmget_not_zero(mm)) {
|
||||
mmap_read_lock(mm);
|
||||
|
||||
/* Re-check after taking map_changing_lock */
|
||||
down_read(&ctx->map_changing_lock);
|
||||
if (likely(!atomic_read(&ctx->mmap_changing)))
|
||||
ret = move_pages(ctx, mm, uffdio_move.dst, uffdio_move.src,
|
||||
uffdio_move.len, uffdio_move.mode);
|
||||
else
|
||||
ret = -EAGAIN;
|
||||
up_read(&ctx->map_changing_lock);
|
||||
mmap_read_unlock(mm);
|
||||
ret = move_pages(ctx, uffdio_move.dst, uffdio_move.src,
|
||||
uffdio_move.len, uffdio_move.mode);
|
||||
mmput(mm);
|
||||
} else {
|
||||
return -ESRCH;
|
||||
|
@ -138,9 +138,8 @@ extern long uffd_wp_range(struct vm_area_struct *vma,
|
||||
/* move_pages */
|
||||
void double_pt_lock(spinlock_t *ptl1, spinlock_t *ptl2);
|
||||
void double_pt_unlock(spinlock_t *ptl1, spinlock_t *ptl2);
|
||||
ssize_t move_pages(struct userfaultfd_ctx *ctx, struct mm_struct *mm,
|
||||
unsigned long dst_start, unsigned long src_start,
|
||||
unsigned long len, __u64 flags);
|
||||
ssize_t move_pages(struct userfaultfd_ctx *ctx, unsigned long dst_start,
|
||||
unsigned long src_start, unsigned long len, __u64 flags);
|
||||
int move_pages_huge_pmd(struct mm_struct *mm, pmd_t *dst_pmd, pmd_t *src_pmd, pmd_t dst_pmdval,
|
||||
struct vm_area_struct *dst_vma,
|
||||
struct vm_area_struct *src_vma,
|
||||
|
@ -2158,7 +2158,7 @@ unlock:
|
||||
|
||||
#ifdef CONFIG_USERFAULTFD
|
||||
/*
|
||||
* The PT lock for src_pmd and the mmap_lock for reading are held by
|
||||
* The PT lock for src_pmd and dst_vma/src_vma (for reading) are locked by
|
||||
* the caller, but it must return after releasing the page_table_lock.
|
||||
* Just move the page from src_pmd to dst_pmd if possible.
|
||||
* Return zero if succeeded in moving the page, -EAGAIN if it needs to be
|
||||
@ -2181,7 +2181,8 @@ int move_pages_huge_pmd(struct mm_struct *mm, pmd_t *dst_pmd, pmd_t *src_pmd, pm
|
||||
src_ptl = pmd_lockptr(mm, src_pmd);
|
||||
|
||||
lockdep_assert_held(src_ptl);
|
||||
mmap_assert_locked(mm);
|
||||
vma_assert_locked(src_vma);
|
||||
vma_assert_locked(dst_vma);
|
||||
|
||||
/* Sanity checks before the operation */
|
||||
if (WARN_ON_ONCE(!pmd_none(dst_pmdval)) || WARN_ON_ONCE(src_addr & ~HPAGE_PMD_MASK) ||
|
||||
|
384
mm/userfaultfd.c
384
mm/userfaultfd.c
@ -20,19 +20,11 @@
|
||||
#include "internal.h"
|
||||
|
||||
static __always_inline
|
||||
struct vm_area_struct *find_dst_vma(struct mm_struct *dst_mm,
|
||||
unsigned long dst_start,
|
||||
unsigned long len)
|
||||
bool validate_dst_vma(struct vm_area_struct *dst_vma, unsigned long dst_end)
|
||||
{
|
||||
/*
|
||||
* Make sure that the dst range is both valid and fully within a
|
||||
* single existing vma.
|
||||
*/
|
||||
struct vm_area_struct *dst_vma;
|
||||
|
||||
dst_vma = find_vma(dst_mm, dst_start);
|
||||
if (!range_in_vma(dst_vma, dst_start, dst_start + len))
|
||||
return NULL;
|
||||
/* Make sure that the dst range is fully within dst_vma. */
|
||||
if (dst_end > dst_vma->vm_end)
|
||||
return false;
|
||||
|
||||
/*
|
||||
* Check the vma is registered in uffd, this is required to
|
||||
@ -40,11 +32,122 @@ struct vm_area_struct *find_dst_vma(struct mm_struct *dst_mm,
|
||||
* time.
|
||||
*/
|
||||
if (!dst_vma->vm_userfaultfd_ctx.ctx)
|
||||
return NULL;
|
||||
return false;
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
static __always_inline
|
||||
struct vm_area_struct *find_vma_and_prepare_anon(struct mm_struct *mm,
|
||||
unsigned long addr)
|
||||
{
|
||||
struct vm_area_struct *vma;
|
||||
|
||||
mmap_assert_locked(mm);
|
||||
vma = vma_lookup(mm, addr);
|
||||
if (!vma)
|
||||
vma = ERR_PTR(-ENOENT);
|
||||
else if (!(vma->vm_flags & VM_SHARED) &&
|
||||
unlikely(anon_vma_prepare(vma)))
|
||||
vma = ERR_PTR(-ENOMEM);
|
||||
|
||||
return vma;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_PER_VMA_LOCK
|
||||
/*
|
||||
* lock_vma() - Lookup and lock vma corresponding to @address.
|
||||
* @mm: mm to search vma in.
|
||||
* @address: address that the vma should contain.
|
||||
*
|
||||
* Should be called without holding mmap_lock. vma should be unlocked after use
|
||||
* with unlock_vma().
|
||||
*
|
||||
* Return: A locked vma containing @address, -ENOENT if no vma is found, or
|
||||
* -ENOMEM if anon_vma couldn't be allocated.
|
||||
*/
|
||||
static struct vm_area_struct *lock_vma(struct mm_struct *mm,
|
||||
unsigned long address)
|
||||
{
|
||||
struct vm_area_struct *vma;
|
||||
|
||||
vma = lock_vma_under_rcu(mm, address);
|
||||
if (vma) {
|
||||
/*
|
||||
* lock_vma_under_rcu() only checks anon_vma for private
|
||||
* anonymous mappings. But we need to ensure it is assigned in
|
||||
* private file-backed vmas as well.
|
||||
*/
|
||||
if (!(vma->vm_flags & VM_SHARED) && unlikely(!vma->anon_vma))
|
||||
vma_end_read(vma);
|
||||
else
|
||||
return vma;
|
||||
}
|
||||
|
||||
mmap_read_lock(mm);
|
||||
vma = find_vma_and_prepare_anon(mm, address);
|
||||
if (!IS_ERR(vma)) {
|
||||
/*
|
||||
* We cannot use vma_start_read() as it may fail due to
|
||||
* false locked (see comment in vma_start_read()). We
|
||||
* can avoid that by directly locking vm_lock under
|
||||
* mmap_lock, which guarantees that nobody can lock the
|
||||
* vma for write (vma_start_write()) under us.
|
||||
*/
|
||||
down_read(&vma->vm_lock->lock);
|
||||
}
|
||||
|
||||
mmap_read_unlock(mm);
|
||||
return vma;
|
||||
}
|
||||
|
||||
static struct vm_area_struct *uffd_mfill_lock(struct mm_struct *dst_mm,
|
||||
unsigned long dst_start,
|
||||
unsigned long len)
|
||||
{
|
||||
struct vm_area_struct *dst_vma;
|
||||
|
||||
dst_vma = lock_vma(dst_mm, dst_start);
|
||||
if (IS_ERR(dst_vma) || validate_dst_vma(dst_vma, dst_start + len))
|
||||
return dst_vma;
|
||||
|
||||
vma_end_read(dst_vma);
|
||||
return ERR_PTR(-ENOENT);
|
||||
}
|
||||
|
||||
static void uffd_mfill_unlock(struct vm_area_struct *vma)
|
||||
{
|
||||
vma_end_read(vma);
|
||||
}
|
||||
|
||||
#else
|
||||
|
||||
static struct vm_area_struct *uffd_mfill_lock(struct mm_struct *dst_mm,
|
||||
unsigned long dst_start,
|
||||
unsigned long len)
|
||||
{
|
||||
struct vm_area_struct *dst_vma;
|
||||
|
||||
mmap_read_lock(dst_mm);
|
||||
dst_vma = find_vma_and_prepare_anon(dst_mm, dst_start);
|
||||
if (IS_ERR(dst_vma))
|
||||
goto out_unlock;
|
||||
|
||||
if (validate_dst_vma(dst_vma, dst_start + len))
|
||||
return dst_vma;
|
||||
|
||||
dst_vma = ERR_PTR(-ENOENT);
|
||||
out_unlock:
|
||||
mmap_read_unlock(dst_mm);
|
||||
return dst_vma;
|
||||
}
|
||||
|
||||
static void uffd_mfill_unlock(struct vm_area_struct *vma)
|
||||
{
|
||||
mmap_read_unlock(vma->vm_mm);
|
||||
}
|
||||
#endif
|
||||
|
||||
/* Check if dst_addr is outside of file's size. Must be called with ptl held. */
|
||||
static bool mfill_file_over_size(struct vm_area_struct *dst_vma,
|
||||
unsigned long dst_addr)
|
||||
@ -350,7 +453,8 @@ static pmd_t *mm_alloc_pmd(struct mm_struct *mm, unsigned long address)
|
||||
#ifdef CONFIG_HUGETLB_PAGE
|
||||
/*
|
||||
* mfill_atomic processing for HUGETLB vmas. Note that this routine is
|
||||
* called with mmap_lock held, it will release mmap_lock before returning.
|
||||
* called with either vma-lock or mmap_lock held, it will release the lock
|
||||
* before returning.
|
||||
*/
|
||||
static __always_inline ssize_t mfill_atomic_hugetlb(
|
||||
struct userfaultfd_ctx *ctx,
|
||||
@ -361,7 +465,6 @@ static __always_inline ssize_t mfill_atomic_hugetlb(
|
||||
uffd_flags_t flags)
|
||||
{
|
||||
struct mm_struct *dst_mm = dst_vma->vm_mm;
|
||||
int vm_shared = dst_vma->vm_flags & VM_SHARED;
|
||||
ssize_t err;
|
||||
pte_t *dst_pte;
|
||||
unsigned long src_addr, dst_addr;
|
||||
@ -380,7 +483,7 @@ static __always_inline ssize_t mfill_atomic_hugetlb(
|
||||
*/
|
||||
if (uffd_flags_mode_is(flags, MFILL_ATOMIC_ZEROPAGE)) {
|
||||
up_read(&ctx->map_changing_lock);
|
||||
mmap_read_unlock(dst_mm);
|
||||
uffd_mfill_unlock(dst_vma);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
@ -403,24 +506,28 @@ retry:
|
||||
* retry, dst_vma will be set to NULL and we must lookup again.
|
||||
*/
|
||||
if (!dst_vma) {
|
||||
dst_vma = uffd_mfill_lock(dst_mm, dst_start, len);
|
||||
if (IS_ERR(dst_vma)) {
|
||||
err = PTR_ERR(dst_vma);
|
||||
goto out;
|
||||
}
|
||||
|
||||
err = -ENOENT;
|
||||
dst_vma = find_dst_vma(dst_mm, dst_start, len);
|
||||
if (!dst_vma || !is_vm_hugetlb_page(dst_vma))
|
||||
goto out_unlock;
|
||||
if (!is_vm_hugetlb_page(dst_vma))
|
||||
goto out_unlock_vma;
|
||||
|
||||
err = -EINVAL;
|
||||
if (vma_hpagesize != vma_kernel_pagesize(dst_vma))
|
||||
goto out_unlock;
|
||||
goto out_unlock_vma;
|
||||
|
||||
vm_shared = dst_vma->vm_flags & VM_SHARED;
|
||||
}
|
||||
|
||||
/*
|
||||
* If not shared, ensure the dst_vma has a anon_vma.
|
||||
*/
|
||||
err = -ENOMEM;
|
||||
if (!vm_shared) {
|
||||
if (unlikely(anon_vma_prepare(dst_vma)))
|
||||
/*
|
||||
* If memory mappings are changing because of non-cooperative
|
||||
* operation (e.g. mremap) running in parallel, bail out and
|
||||
* request the user to retry later
|
||||
*/
|
||||
down_read(&ctx->map_changing_lock);
|
||||
err = -EAGAIN;
|
||||
if (atomic_read(&ctx->mmap_changing))
|
||||
goto out_unlock;
|
||||
}
|
||||
|
||||
@ -465,7 +572,7 @@ retry:
|
||||
|
||||
if (unlikely(err == -ENOENT)) {
|
||||
up_read(&ctx->map_changing_lock);
|
||||
mmap_read_unlock(dst_mm);
|
||||
uffd_mfill_unlock(dst_vma);
|
||||
BUG_ON(!folio);
|
||||
|
||||
err = copy_folio_from_user(folio,
|
||||
@ -474,17 +581,6 @@ retry:
|
||||
err = -EFAULT;
|
||||
goto out;
|
||||
}
|
||||
mmap_read_lock(dst_mm);
|
||||
down_read(&ctx->map_changing_lock);
|
||||
/*
|
||||
* If memory mappings are changing because of non-cooperative
|
||||
* operation (e.g. mremap) running in parallel, bail out and
|
||||
* request the user to retry later
|
||||
*/
|
||||
if (atomic_read(&ctx->mmap_changing)) {
|
||||
err = -EAGAIN;
|
||||
break;
|
||||
}
|
||||
|
||||
dst_vma = NULL;
|
||||
goto retry;
|
||||
@ -505,7 +601,8 @@ retry:
|
||||
|
||||
out_unlock:
|
||||
up_read(&ctx->map_changing_lock);
|
||||
mmap_read_unlock(dst_mm);
|
||||
out_unlock_vma:
|
||||
uffd_mfill_unlock(dst_vma);
|
||||
out:
|
||||
if (folio)
|
||||
folio_put(folio);
|
||||
@ -597,7 +694,15 @@ static __always_inline ssize_t mfill_atomic(struct userfaultfd_ctx *ctx,
|
||||
copied = 0;
|
||||
folio = NULL;
|
||||
retry:
|
||||
mmap_read_lock(dst_mm);
|
||||
/*
|
||||
* Make sure the vma is not shared, that the dst range is
|
||||
* both valid and fully within a single existing vma.
|
||||
*/
|
||||
dst_vma = uffd_mfill_lock(dst_mm, dst_start, len);
|
||||
if (IS_ERR(dst_vma)) {
|
||||
err = PTR_ERR(dst_vma);
|
||||
goto out;
|
||||
}
|
||||
|
||||
/*
|
||||
* If memory mappings are changing because of non-cooperative
|
||||
@ -609,15 +714,6 @@ retry:
|
||||
if (atomic_read(&ctx->mmap_changing))
|
||||
goto out_unlock;
|
||||
|
||||
/*
|
||||
* Make sure the vma is not shared, that the dst range is
|
||||
* both valid and fully within a single existing vma.
|
||||
*/
|
||||
err = -ENOENT;
|
||||
dst_vma = find_dst_vma(dst_mm, dst_start, len);
|
||||
if (!dst_vma)
|
||||
goto out_unlock;
|
||||
|
||||
err = -EINVAL;
|
||||
/*
|
||||
* shmem_zero_setup is invoked in mmap for MAP_ANONYMOUS|MAP_SHARED but
|
||||
@ -647,16 +743,6 @@ retry:
|
||||
uffd_flags_mode_is(flags, MFILL_ATOMIC_CONTINUE))
|
||||
goto out_unlock;
|
||||
|
||||
/*
|
||||
* Ensure the dst_vma has a anon_vma or this page
|
||||
* would get a NULL anon_vma when moved in the
|
||||
* dst_vma.
|
||||
*/
|
||||
err = -ENOMEM;
|
||||
if (!(dst_vma->vm_flags & VM_SHARED) &&
|
||||
unlikely(anon_vma_prepare(dst_vma)))
|
||||
goto out_unlock;
|
||||
|
||||
while (src_addr < src_start + len) {
|
||||
pmd_t dst_pmdval;
|
||||
|
||||
@ -699,7 +785,7 @@ retry:
|
||||
void *kaddr;
|
||||
|
||||
up_read(&ctx->map_changing_lock);
|
||||
mmap_read_unlock(dst_mm);
|
||||
uffd_mfill_unlock(dst_vma);
|
||||
BUG_ON(!folio);
|
||||
|
||||
kaddr = kmap_local_folio(folio, 0);
|
||||
@ -730,7 +816,7 @@ retry:
|
||||
|
||||
out_unlock:
|
||||
up_read(&ctx->map_changing_lock);
|
||||
mmap_read_unlock(dst_mm);
|
||||
uffd_mfill_unlock(dst_vma);
|
||||
out:
|
||||
if (folio)
|
||||
folio_put(folio);
|
||||
@ -1267,27 +1353,136 @@ static int validate_move_areas(struct userfaultfd_ctx *ctx,
|
||||
if (!vma_is_anonymous(src_vma) || !vma_is_anonymous(dst_vma))
|
||||
return -EINVAL;
|
||||
|
||||
/*
|
||||
* Ensure the dst_vma has a anon_vma or this page
|
||||
* would get a NULL anon_vma when moved in the
|
||||
* dst_vma.
|
||||
*/
|
||||
if (unlikely(anon_vma_prepare(dst_vma)))
|
||||
return -ENOMEM;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static __always_inline
|
||||
int find_vmas_mm_locked(struct mm_struct *mm,
|
||||
unsigned long dst_start,
|
||||
unsigned long src_start,
|
||||
struct vm_area_struct **dst_vmap,
|
||||
struct vm_area_struct **src_vmap)
|
||||
{
|
||||
struct vm_area_struct *vma;
|
||||
|
||||
mmap_assert_locked(mm);
|
||||
vma = find_vma_and_prepare_anon(mm, dst_start);
|
||||
if (IS_ERR(vma))
|
||||
return PTR_ERR(vma);
|
||||
|
||||
*dst_vmap = vma;
|
||||
/* Skip finding src_vma if src_start is in dst_vma */
|
||||
if (src_start >= vma->vm_start && src_start < vma->vm_end)
|
||||
goto out_success;
|
||||
|
||||
vma = vma_lookup(mm, src_start);
|
||||
if (!vma)
|
||||
return -ENOENT;
|
||||
out_success:
|
||||
*src_vmap = vma;
|
||||
return 0;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_PER_VMA_LOCK
|
||||
static int uffd_move_lock(struct mm_struct *mm,
|
||||
unsigned long dst_start,
|
||||
unsigned long src_start,
|
||||
struct vm_area_struct **dst_vmap,
|
||||
struct vm_area_struct **src_vmap)
|
||||
{
|
||||
struct vm_area_struct *vma;
|
||||
int err;
|
||||
|
||||
vma = lock_vma(mm, dst_start);
|
||||
if (IS_ERR(vma))
|
||||
return PTR_ERR(vma);
|
||||
|
||||
*dst_vmap = vma;
|
||||
/*
|
||||
* Skip finding src_vma if src_start is in dst_vma. This also ensures
|
||||
* that we don't lock the same vma twice.
|
||||
*/
|
||||
if (src_start >= vma->vm_start && src_start < vma->vm_end) {
|
||||
*src_vmap = vma;
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Using lock_vma() to get src_vma can lead to following deadlock:
|
||||
*
|
||||
* Thread1 Thread2
|
||||
* ------- -------
|
||||
* vma_start_read(dst_vma)
|
||||
* mmap_write_lock(mm)
|
||||
* vma_start_write(src_vma)
|
||||
* vma_start_read(src_vma)
|
||||
* mmap_read_lock(mm)
|
||||
* vma_start_write(dst_vma)
|
||||
*/
|
||||
*src_vmap = lock_vma_under_rcu(mm, src_start);
|
||||
if (likely(*src_vmap))
|
||||
return 0;
|
||||
|
||||
/* Undo any locking and retry in mmap_lock critical section */
|
||||
vma_end_read(*dst_vmap);
|
||||
|
||||
mmap_read_lock(mm);
|
||||
err = find_vmas_mm_locked(mm, dst_start, src_start, dst_vmap, src_vmap);
|
||||
if (!err) {
|
||||
/*
|
||||
* See comment in lock_vma() as to why not using
|
||||
* vma_start_read() here.
|
||||
*/
|
||||
down_read(&(*dst_vmap)->vm_lock->lock);
|
||||
if (*dst_vmap != *src_vmap)
|
||||
down_read(&(*src_vmap)->vm_lock->lock);
|
||||
}
|
||||
mmap_read_unlock(mm);
|
||||
return err;
|
||||
}
|
||||
|
||||
static void uffd_move_unlock(struct vm_area_struct *dst_vma,
|
||||
struct vm_area_struct *src_vma)
|
||||
{
|
||||
vma_end_read(src_vma);
|
||||
if (src_vma != dst_vma)
|
||||
vma_end_read(dst_vma);
|
||||
}
|
||||
|
||||
#else
|
||||
|
||||
static int uffd_move_lock(struct mm_struct *mm,
|
||||
unsigned long dst_start,
|
||||
unsigned long src_start,
|
||||
struct vm_area_struct **dst_vmap,
|
||||
struct vm_area_struct **src_vmap)
|
||||
{
|
||||
int err;
|
||||
|
||||
mmap_read_lock(mm);
|
||||
err = find_vmas_mm_locked(mm, dst_start, src_start, dst_vmap, src_vmap);
|
||||
if (err)
|
||||
mmap_read_unlock(mm);
|
||||
return err;
|
||||
}
|
||||
|
||||
static void uffd_move_unlock(struct vm_area_struct *dst_vma,
|
||||
struct vm_area_struct *src_vma)
|
||||
{
|
||||
mmap_assert_locked(src_vma->vm_mm);
|
||||
mmap_read_unlock(dst_vma->vm_mm);
|
||||
}
|
||||
#endif
|
||||
|
||||
/**
|
||||
* move_pages - move arbitrary anonymous pages of an existing vma
|
||||
* @ctx: pointer to the userfaultfd context
|
||||
* @mm: the address space to move pages
|
||||
* @dst_start: start of the destination virtual memory range
|
||||
* @src_start: start of the source virtual memory range
|
||||
* @len: length of the virtual memory range
|
||||
* @mode: flags from uffdio_move.mode
|
||||
*
|
||||
* Must be called with mmap_lock held for read.
|
||||
* It will either use the mmap_lock in read mode or per-vma locks
|
||||
*
|
||||
* move_pages() remaps arbitrary anonymous pages atomically in zero
|
||||
* copy. It only works on non shared anonymous pages because those can
|
||||
@ -1355,10 +1550,10 @@ static int validate_move_areas(struct userfaultfd_ctx *ctx,
|
||||
* could be obtained. This is the only additional complexity added to
|
||||
* the rmap code to provide this anonymous page remapping functionality.
|
||||
*/
|
||||
ssize_t move_pages(struct userfaultfd_ctx *ctx, struct mm_struct *mm,
|
||||
unsigned long dst_start, unsigned long src_start,
|
||||
unsigned long len, __u64 mode)
|
||||
ssize_t move_pages(struct userfaultfd_ctx *ctx, unsigned long dst_start,
|
||||
unsigned long src_start, unsigned long len, __u64 mode)
|
||||
{
|
||||
struct mm_struct *mm = ctx->mm;
|
||||
struct vm_area_struct *src_vma, *dst_vma;
|
||||
unsigned long src_addr, dst_addr;
|
||||
pmd_t *src_pmd, *dst_pmd;
|
||||
@ -1376,28 +1571,34 @@ ssize_t move_pages(struct userfaultfd_ctx *ctx, struct mm_struct *mm,
|
||||
WARN_ON_ONCE(dst_start + len <= dst_start))
|
||||
goto out;
|
||||
|
||||
err = uffd_move_lock(mm, dst_start, src_start, &dst_vma, &src_vma);
|
||||
if (err)
|
||||
goto out;
|
||||
|
||||
/* Re-check after taking map_changing_lock */
|
||||
err = -EAGAIN;
|
||||
down_read(&ctx->map_changing_lock);
|
||||
if (likely(atomic_read(&ctx->mmap_changing)))
|
||||
goto out_unlock;
|
||||
/*
|
||||
* Make sure the vma is not shared, that the src and dst remap
|
||||
* ranges are both valid and fully within a single existing
|
||||
* vma.
|
||||
*/
|
||||
src_vma = find_vma(mm, src_start);
|
||||
if (!src_vma || (src_vma->vm_flags & VM_SHARED))
|
||||
goto out;
|
||||
if (src_start < src_vma->vm_start ||
|
||||
src_start + len > src_vma->vm_end)
|
||||
goto out;
|
||||
err = -EINVAL;
|
||||
if (src_vma->vm_flags & VM_SHARED)
|
||||
goto out_unlock;
|
||||
if (src_start + len > src_vma->vm_end)
|
||||
goto out_unlock;
|
||||
|
||||
dst_vma = find_vma(mm, dst_start);
|
||||
if (!dst_vma || (dst_vma->vm_flags & VM_SHARED))
|
||||
goto out;
|
||||
if (dst_start < dst_vma->vm_start ||
|
||||
dst_start + len > dst_vma->vm_end)
|
||||
goto out;
|
||||
if (dst_vma->vm_flags & VM_SHARED)
|
||||
goto out_unlock;
|
||||
if (dst_start + len > dst_vma->vm_end)
|
||||
goto out_unlock;
|
||||
|
||||
err = validate_move_areas(ctx, src_vma, dst_vma);
|
||||
if (err)
|
||||
goto out;
|
||||
goto out_unlock;
|
||||
|
||||
for (src_addr = src_start, dst_addr = dst_start;
|
||||
src_addr < src_start + len;) {
|
||||
@ -1514,6 +1715,9 @@ ssize_t move_pages(struct userfaultfd_ctx *ctx, struct mm_struct *mm,
|
||||
moved += step_size;
|
||||
}
|
||||
|
||||
out_unlock:
|
||||
up_read(&ctx->map_changing_lock);
|
||||
uffd_move_unlock(dst_vma, src_vma);
|
||||
out:
|
||||
VM_WARN_ON(moved < 0);
|
||||
VM_WARN_ON(err > 0);
|
||||
|
Loading…
Reference in New Issue
Block a user