mm: remove get_swap_bio
[linux-2.6-microblaze.git] / mm / swapfile.c
index 1c0a829..bfa9e8b 100644 (file)
@@ -1042,16 +1042,18 @@ int get_swap_pages(int n_goal, swp_entry_t swp_entries[], int entry_size)
        /* Only single cluster request supported */
        WARN_ON_ONCE(n_goal > 1 && size == SWAPFILE_CLUSTER);
 
+       spin_lock(&swap_avail_lock);
+
        avail_pgs = atomic_long_read(&nr_swap_pages) / size;
-       if (avail_pgs <= 0)
+       if (avail_pgs <= 0) {
+               spin_unlock(&swap_avail_lock);
                goto noswap;
+       }
 
        n_goal = min3((long)n_goal, (long)SWAP_BATCH, avail_pgs);
 
        atomic_long_sub(n_goal * size, &nr_swap_pages);
 
-       spin_lock(&swap_avail_lock);
-
 start_over:
        node = numa_node_id();
        plist_for_each_entry_safe(si, next, &swap_avail_heads[node], avail_lists[node]) {
@@ -1125,14 +1127,13 @@ swp_entry_t get_swap_page_of_type(int type)
 
        spin_lock(&si->lock);
        if (si->flags & SWP_WRITEOK) {
-               atomic_long_dec(&nr_swap_pages);
                /* This is called for allocating swap entry, not cache */
                offset = scan_swap_map(si, 1);
                if (offset) {
+                       atomic_long_dec(&nr_swap_pages);
                        spin_unlock(&si->lock);
                        return swp_entry(type, offset);
                }
-               atomic_long_inc(&nr_swap_pages);
        }
        spin_unlock(&si->lock);
 fail:
@@ -2300,16 +2301,6 @@ static sector_t map_swap_entry(swp_entry_t entry, struct block_device **bdev)
        return se->start_block + (offset - se->start_page);
 }
 
-/*
- * Returns the page offset into bdev for the specified page's swap entry.
- */
-sector_t map_swap_page(struct page *page, struct block_device **bdev)
-{
-       swp_entry_t entry;
-       entry.val = page_private(page);
-       return map_swap_entry(entry, bdev);
-}
-
 /*
  * Free all of a swapdev's extent information
  */