mm/mmap: not necessary to check mapping separately
[linux-2.6-microblaze.git] / mm / mmap.c
index f793eb7..295197b 100644 (file)
--- a/mm/mmap.c
+++ b/mm/mmap.c
@@ -474,8 +474,12 @@ static __always_inline void vma_rb_erase_ignore(struct vm_area_struct *vma,
 {
        /*
         * All rb_subtree_gap values must be consistent prior to erase,
-        * with the possible exception of the "next" vma being erased if
-        * next->vm_start was reduced.
+        * with the possible exception of
+        *
+        * a. the "next" vma being erased if next->vm_start was reduced in
+        *    __vma_adjust() -> __vma_unlink()
+        * b. the vma being erased in detach_vmas_to_be_unmapped() ->
+        *    vma_rb_erase()
         */
        validate_mm_rb(root, ignore);
 
@@ -485,13 +489,7 @@ static __always_inline void vma_rb_erase_ignore(struct vm_area_struct *vma,
 static __always_inline void vma_rb_erase(struct vm_area_struct *vma,
                                         struct rb_root *root)
 {
-       /*
-        * All rb_subtree_gap values must be consistent prior to erase,
-        * with the possible exception of the vma being erased.
-        */
-       validate_mm_rb(root, vma);
-
-       __vma_rb_erase(vma, root);
+       vma_rb_erase_ignore(vma, root, vma);
 }
 
 /*
@@ -677,7 +675,7 @@ static void __insert_vm_struct(struct mm_struct *mm, struct vm_area_struct *vma)
        mm->map_count++;
 }
 
-static __always_inline void __vma_unlink_common(struct mm_struct *mm,
+static __always_inline void __vma_unlink(struct mm_struct *mm,
                                                struct vm_area_struct *vma,
                                                struct vm_area_struct *ignore)
 {
@@ -760,7 +758,7 @@ int __vma_adjust(struct vm_area_struct *vma, unsigned long start,
                         * vma expands, overlapping part of the next:
                         * mprotect case 5 shifting the boundary up.
                         */
-                       adjust_next = (end - next->vm_start) >> PAGE_SHIFT;
+                       adjust_next = (end - next->vm_start);
                        exporter = next;
                        importer = vma;
                        VM_WARN_ON(expand != importer);
@@ -770,7 +768,7 @@ int __vma_adjust(struct vm_area_struct *vma, unsigned long start,
                         * split_vma inserting another: so it must be
                         * mprotect case 4 shifting the boundary down.
                         */
-                       adjust_next = -((vma->vm_end - end) >> PAGE_SHIFT);
+                       adjust_next = -(vma->vm_end - end);
                        exporter = vma;
                        importer = next;
                        VM_WARN_ON(expand != importer);
@@ -842,8 +840,8 @@ again:
        }
        vma->vm_pgoff = pgoff;
        if (adjust_next) {
-               next->vm_start += adjust_next << PAGE_SHIFT;
-               next->vm_pgoff += adjust_next;
+               next->vm_start += adjust_next;
+               next->vm_pgoff += adjust_next >> PAGE_SHIFT;
        }
 
        if (root) {
@@ -859,7 +857,7 @@ again:
                 * us to remove next before dropping the locks.
                 */
                if (remove_next != 3)
-                       __vma_unlink_common(mm, next, next);
+                       __vma_unlink(mm, next, next);
                else
                        /*
                         * vma is not before next if they've been
@@ -870,7 +868,7 @@ again:
                         * "next" (which is stored in post-swap()
                         * "vma").
                         */
-                       __vma_unlink_common(mm, next, vma);
+                       __vma_unlink(mm, next, vma);
                if (file)
                        __remove_shared_vm_struct(next, file, mapping);
        } else if (insert) {
@@ -897,10 +895,9 @@ again:
                        anon_vma_interval_tree_post_update_vma(next);
                anon_vma_unlock_write(anon_vma);
        }
-       if (mapping)
-               i_mmap_unlock_write(mapping);
 
        if (root) {
+               i_mmap_unlock_write(mapping);
                uprobe_mmap(vma);
 
                if (adjust_next)
@@ -1666,7 +1663,7 @@ int vma_wants_writenotify(struct vm_area_struct *vma, pgprot_t vm_page_prot)
 
        /* Can the mapping track the dirty pages? */
        return vma->vm_file && vma->vm_file->f_mapping &&
-               mapping_cap_account_dirty(vma->vm_file->f_mapping);
+               mapping_can_writeback(vma->vm_file->f_mapping);
 }
 
 /*