diff mbox series

[mm-unstable,1/2] mm/mm_init: rename __init_reserved_page_zone to __init_page_from_nid

Message ID 20250225083017.567649-2-rppt@kernel.org (mailing list archive)
State New
Headers show
Series mm_init: rename *init_reserved_page* functions | expand

Commit Message

Mike Rapoport Feb. 25, 2025, 8:30 a.m. UTC
From: "Mike Rapoport (Microsoft)" <rppt@kernel.org>

__init_reserved_page_zone() function finds the zone for pfn and nid and
performs initialization of a struct page with that zone and nid.
There is nothing in that function about reserved pages and it is
misnamed.

Rename it to __init_page_from_nid() to better reflect what the function
does.

Signed-off-by: Mike Rapoport (Microsoft) <rppt@kernel.org>
---
 mm/hugetlb.c  | 2 +-
 mm/internal.h | 2 +-
 mm/mm_init.c  | 4 ++--
 3 files changed, 4 insertions(+), 4 deletions(-)

Comments

Wei Yang Feb. 26, 2025, 1:52 a.m. UTC | #1
On Tue, Feb 25, 2025 at 10:30:16AM +0200, Mike Rapoport wrote:
>From: "Mike Rapoport (Microsoft)" <rppt@kernel.org>
>
>__init_reserved_page_zone() function finds the zone for pfn and nid and
>performs initialization of a struct page with that zone and nid.
>There is nothing in that function about reserved pages and it is
>misnamed.
>
>Rename it to __init_page_from_nid() to better reflect what the function
>does.
>
>Signed-off-by: Mike Rapoport (Microsoft) <rppt@kernel.org>

Reviewed-by: Wei Yang <richard.weiyang@gmail.com>
diff mbox series

Patch

diff --git a/mm/hugetlb.c b/mm/hugetlb.c
index 9faa1034704f..778c21da22f0 100644
--- a/mm/hugetlb.c
+++ b/mm/hugetlb.c
@@ -3392,7 +3392,7 @@  static void __init hugetlb_bootmem_free_invalid_page(int nid, struct page *page,
 
 	while (npages--) {
 		pfn = page_to_pfn(page);
-		__init_reserved_page_zone(pfn, nid);
+		__init_page_from_nid(pfn, nid);
 		free_reserved_page(page);
 		page++;
 	}
diff --git a/mm/internal.h b/mm/internal.h
index b07550db2bfd..5c0b81abbc1b 100644
--- a/mm/internal.h
+++ b/mm/internal.h
@@ -1465,7 +1465,7 @@  static inline bool pte_needs_soft_dirty_wp(struct vm_area_struct *vma, pte_t pte
 
 void __meminit __init_single_page(struct page *page, unsigned long pfn,
 				unsigned long zone, int nid);
-void __meminit __init_reserved_page_zone(unsigned long pfn, int nid);
+void __meminit __init_page_from_nid(unsigned long pfn, int nid);
 
 /* shrinker related functions */
 unsigned long shrink_slab(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg,
diff --git a/mm/mm_init.c b/mm/mm_init.c
index 133640a93d1d..f9a02b1e2fce 100644
--- a/mm/mm_init.c
+++ b/mm/mm_init.c
@@ -653,7 +653,7 @@  static inline void fixup_hashdist(void) {}
 /*
  * Initialize a reserved page unconditionally, finding its zone first.
  */
-void __meminit __init_reserved_page_zone(unsigned long pfn, int nid)
+void __meminit __init_page_from_nid(unsigned long pfn, int nid)
 {
 	pg_data_t *pgdat;
 	int zid;
@@ -733,7 +733,7 @@  static void __meminit init_reserved_page(unsigned long pfn, int nid)
 	if (early_page_initialised(pfn, nid))
 		return;
 
-	__init_reserved_page_zone(pfn, nid);
+	__init_page_from_nid(pfn, nid);
 }
 #else
 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) {}