mm/madvise: clean up force_shm_swapin_readahead()
authorHugh Dickins <hughd@google.com>
Fri, 9 Jun 2023 01:35:14 +0000 (18:35 -0700)
committerAndrew Morton <akpm@linux-foundation.org>
Mon, 19 Jun 2023 23:19:16 +0000 (16:19 -0700)
Some nearby MADV_WILLNEED cleanup unrelated to pte_offset_map_lock().
shmem_swapin_range() is a better name than force_shm_swapin_readahead().
Fix unimportant off-by-one on end_index.  Call the swp_entry_t "entry"
rather than "swap": either is okay, but entry is the name used elsewhere
in mm/madvise.c.  Do not assume GFP_HIGHUSER_MOVABLE: that's right for
anon swap, but shmem should take gfp from mapping.  Pass the actual vma
and address to read_swap_cache_async(), in case a NUMA mempolicy applies.
lru_add_drain() at outer level, like madvise_willneed()'s other branch.

Link: https://lkml.kernel.org/r/67e18875-ffb3-ec27-346-f350e07bed87@google.com
Signed-off-by: Hugh Dickins <hughd@google.com>
Cc: Alistair Popple <apopple@nvidia.com>
Cc: Anshuman Khandual <anshuman.khandual@arm.com>
Cc: Axel Rasmussen <axelrasmussen@google.com>
Cc: Christophe Leroy <christophe.leroy@csgroup.eu>
Cc: Christoph Hellwig <hch@infradead.org>
Cc: David Hildenbrand <david@redhat.com>
Cc: "Huang, Ying" <ying.huang@intel.com>
Cc: Ira Weiny <ira.weiny@intel.com>
Cc: Jason Gunthorpe <jgg@ziepe.ca>
Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Cc: Lorenzo Stoakes <lstoakes@gmail.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Mel Gorman <mgorman@techsingularity.net>
Cc: Miaohe Lin <linmiaohe@huawei.com>
Cc: Mike Kravetz <mike.kravetz@oracle.com>
Cc: Mike Rapoport (IBM) <rppt@kernel.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Naoya Horiguchi <naoya.horiguchi@nec.com>
Cc: Pavel Tatashin <pasha.tatashin@soleen.com>
Cc: Peter Xu <peterx@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Qi Zheng <zhengqi.arch@bytedance.com>
Cc: Ralph Campbell <rcampbell@nvidia.com>
Cc: Ryan Roberts <ryan.roberts@arm.com>
Cc: SeongJae Park <sj@kernel.org>
Cc: Song Liu <song@kernel.org>
Cc: Steven Price <steven.price@arm.com>
Cc: Suren Baghdasaryan <surenb@google.com>
Cc: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Cc: Will Deacon <will@kernel.org>
Cc: Yang Shi <shy828301@gmail.com>
Cc: Yu Zhao <yuzhao@google.com>
Cc: Zack Rusin <zackr@vmware.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
mm/madvise.c

index 0af64c4..9b3c961 100644 (file)
@@ -235,30 +235,34 @@ static const struct mm_walk_ops swapin_walk_ops = {
        .pmd_entry              = swapin_walk_pmd_entry,
 };
 
-static void force_shm_swapin_readahead(struct vm_area_struct *vma,
+static void shmem_swapin_range(struct vm_area_struct *vma,
                unsigned long start, unsigned long end,
                struct address_space *mapping)
 {
        XA_STATE(xas, &mapping->i_pages, linear_page_index(vma, start));
-       pgoff_t end_index = linear_page_index(vma, end + PAGE_SIZE - 1);
+       pgoff_t end_index = linear_page_index(vma, end) - 1;
        struct page *page;
        struct swap_iocb *splug = NULL;
 
        rcu_read_lock();
        xas_for_each(&xas, page, end_index) {
-               swp_entry_t swap;
+               unsigned long addr;
+               swp_entry_t entry;
 
                if (!xa_is_value(page))
                        continue;
-               swap = radix_to_swp_entry(page);
+               entry = radix_to_swp_entry(page);
                /* There might be swapin error entries in shmem mapping. */
-               if (non_swap_entry(swap))
+               if (non_swap_entry(entry))
                        continue;
+
+               addr = vma->vm_start +
+                       ((xas.xa_index - vma->vm_pgoff) << PAGE_SHIFT);
                xas_pause(&xas);
                rcu_read_unlock();
 
-               page = read_swap_cache_async(swap, GFP_HIGHUSER_MOVABLE,
-                                            NULL, 0, false, &splug);
+               page = read_swap_cache_async(entry, mapping_gfp_mask(mapping),
+                                            vma, addr, false, &splug);
                if (page)
                        put_page(page);
 
@@ -266,8 +270,6 @@ static void force_shm_swapin_readahead(struct vm_area_struct *vma,
        }
        rcu_read_unlock();
        swap_read_unplug(splug);
-
-       lru_add_drain();        /* Push any new pages onto the LRU now */
 }
 #endif         /* CONFIG_SWAP */
 
@@ -291,8 +293,8 @@ static long madvise_willneed(struct vm_area_struct *vma,
        }
 
        if (shmem_mapping(file->f_mapping)) {
-               force_shm_swapin_readahead(vma, start, end,
-                                       file->f_mapping);
+               shmem_swapin_range(vma, start, end, file->f_mapping);
+               lru_add_drain(); /* Push any new pages onto the LRU now */
                return 0;
        }
 #else