mm: convert free_pages_and_swap_cache() to use folios_put()
Process the pages in batch-sized quantities instead of all-at-once. Link: https://lkml.kernel.org/r/20240227174254.710559-17-willy@infradead.org Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org> Cc: David Hildenbrand <david@redhat.com> Cc: Mel Gorman <mgorman@suse.de> Cc: Ryan Roberts <ryan.roberts@arm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
f39ec4dcb9
commit
4907e80b76
@ -15,6 +15,7 @@
|
||||
#include <linux/swapops.h>
|
||||
#include <linux/init.h>
|
||||
#include <linux/pagemap.h>
|
||||
#include <linux/pagevec.h>
|
||||
#include <linux/backing-dev.h>
|
||||
#include <linux/blkdev.h>
|
||||
#include <linux/migrate.h>
|
||||
@ -310,21 +311,25 @@ void free_page_and_swap_cache(struct page *page)
|
||||
*/
|
||||
void free_pages_and_swap_cache(struct encoded_page **pages, int nr)
|
||||
{
|
||||
lru_add_drain();
|
||||
for (int i = 0; i < nr; i++) {
|
||||
struct page *page = encoded_page_ptr(pages[i]);
|
||||
struct folio_batch folios;
|
||||
unsigned int refs[PAGEVEC_SIZE];
|
||||
|
||||
/*
|
||||
* Skip over the "nr_pages" entry. It's sufficient to call
|
||||
* free_swap_cache() only once per folio.
|
||||
*/
|
||||
lru_add_drain();
|
||||
folio_batch_init(&folios);
|
||||
for (int i = 0; i < nr; i++) {
|
||||
struct folio *folio = page_folio(encoded_page_ptr(pages[i]));
|
||||
|
||||
free_swap_cache(&folio->page);
|
||||
refs[folios.nr] = 1;
|
||||
if (unlikely(encoded_page_flags(pages[i]) &
|
||||
ENCODED_PAGE_BIT_NR_PAGES_NEXT))
|
||||
i++;
|
||||
refs[folios.nr] = encoded_nr_pages(pages[++i]);
|
||||
|
||||
free_swap_cache(page);
|
||||
if (folio_batch_add(&folios, folio) == 0)
|
||||
folios_put_refs(&folios, refs);
|
||||
}
|
||||
release_pages(pages, nr);
|
||||
if (folios.nr)
|
||||
folios_put_refs(&folios, refs);
|
||||
}
|
||||
|
||||
static inline bool swap_use_vma_readahead(void)
|
||||
|
Loading…
Reference in New Issue
Block a user