@@ -680,6 +680,7 @@ extern int user_min_free_kbytes;
void free_unref_page(struct page *page, unsigned int order);
void free_unref_folios(struct folio_batch *fbatch);
+void free_unref_page_list(struct list_head *list);
extern void zone_pcp_reset(struct zone *zone);
extern void zone_pcp_disable(struct zone *zone);
@@ -2774,6 +2774,24 @@ void free_unref_folios(struct folio_batch *folios)
folio_batch_reinit(folios);
}
+void free_unref_page_list(struct list_head *list)
+{
+ struct folio_batch fbatch;
+
+ folio_batch_init(&fbatch);
+ while (!list_empty(list)) {
+ struct folio *folio = list_first_entry(list, struct folio, lru);
+
+ list_del(&folio->lru);
+ if (folio_batch_add(&fbatch, folio) > 0)
+ continue;
+ free_unref_folios(&fbatch);
+ }
+
+ if (fbatch.nr)
+ free_unref_folios(&fbatch);
+}
+
/*
* split_page takes a non-compound higher-order page, and splits it into
* n (1<<order) sub-pages: page[0..n]
free_unref_page_list will be needed in a later patch for an optimization to free zapped tail pages when splitting isolated thp. Signed-off-by: Usama Arif <usamaarif642@gmail.com> --- mm/internal.h | 1 + mm/page_alloc.c | 18 ++++++++++++++++++ 2 files changed, 19 insertions(+)