From patchwork Wed Feb 26 12:01:29 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Byungchul Park X-Patchwork-Id: 13992194 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 973A3C021B8 for ; Wed, 26 Feb 2025 12:02:48 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id B932B280033; Wed, 26 Feb 2025 07:02:08 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id B448228002D; Wed, 26 Feb 2025 07:02:08 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 9980F280033; Wed, 26 Feb 2025 07:02:08 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0011.hostedemail.com [216.40.44.11]) by kanga.kvack.org (Postfix) with ESMTP id 414B628002D for ; Wed, 26 Feb 2025 07:02:08 -0500 (EST) Received: from smtpin27.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay10.hostedemail.com (Postfix) with ESMTP id 4CE9EC13FD for ; Wed, 26 Feb 2025 12:01:56 +0000 (UTC) X-FDA: 83161957032.27.96FA760 Received: from invmail4.hynix.com (exvmail4.skhynix.com [166.125.252.92]) by imf16.hostedemail.com (Postfix) with ESMTP id 00059180034 for ; Wed, 26 Feb 2025 12:01:52 +0000 (UTC) Authentication-Results: imf16.hostedemail.com; dkim=none; dmarc=none; spf=pass (imf16.hostedemail.com: domain of byungchul@sk.com designates 166.125.252.92 as permitted sender) smtp.mailfrom=byungchul@sk.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1740571313; a=rsa-sha256; cv=none; b=OHUSIiqZHPPe/soCvB1ZYUlVy0R9qBtIXzIw0qdNxzby9T+r6QzrRxW7CS7t79Qp2RjLcR bXXi2LnkhJc9d1mXEKYl/VPRqJ9X8OfDhY6ymMPSRdKGyniXbUHmVZ19QGH/ALDBs0UTk9 EDEl0XdkS0XgLg7lcRIwVUHuT0/tsDE= ARC-Authentication-Results: i=1; imf16.hostedemail.com; dkim=none; dmarc=none; spf=pass (imf16.hostedemail.com: domain of byungchul@sk.com designates 166.125.252.92 as permitted sender) smtp.mailfrom=byungchul@sk.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1740571313; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:content-type: content-transfer-encoding:in-reply-to:in-reply-to: references:references; bh=yt/j7uLHHbHD82MNbKO0fiJJBF3HG+noFKUKLwwlwIo=; b=7iXun15eLQWZWAGEiNGQbczklAjyZnAgg6MQU+F/M3qK37HvShVwbcVTzR9w2tyEPKxBYk UOgljlp/7j/zcPOlf9C4dtPAQe9ObFociM5+GUcOAGgImM/jKq8HnOloz+v+lbAlmnAeTk kfrfdsoIap59fR/J5mWmOcXTUqk+6nw= X-AuditID: a67dfc5b-3e1ff7000001d7ae-31-67bf02a743c9 From: Byungchul Park To: linux-kernel@vger.kernel.org, linux-mm@kvack.org Cc: kernel_team@skhynix.com, akpm@linux-foundation.org, vernhao@tencent.com, mgorman@techsingularity.net, hughd@google.com, willy@infradead.org, david@redhat.com, peterz@infradead.org, luto@kernel.org, tglx@linutronix.de, mingo@redhat.com, bp@alien8.de, rjgolo@gmail.com Subject: [RFC PATCH v12 based on mm-unstable as of Feb 21, 2025 22/25] mm/page_alloc: not allow to tlb shootdown if !preemptable() && non_luf_pages_ok() Date: Wed, 26 Feb 2025 21:01:29 +0900 Message-Id: <20250226120132.28469-22-byungchul@sk.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20250226120132.28469-1-byungchul@sk.com> References: <20250226113342.GB1935@system.software.com> <20250226120132.28469-1-byungchul@sk.com> X-Brightmail-Tracker: H4sIAAAAAAAAA+NgFrrKLMWRmVeSWpSXmKPExsXC9ZZnke5ypv3pBj0TtS3mrF/DZvF5wz82 i6/rfzFbPP3Ux2JxedccNot7a/6zWpzftZbVYsfSfUwWlw4sYLI43nuAyWL+vc9sFps3TWW2 OD5lKqPF7x9z2Bz4PL639rF47Jx1l91jwaZSj80rtDw2repk89j0aRK7x7tz59g9Tsz4zeLx ft9VNo+tv+w8GqdeY/P4vEkugCeKyyYlNSezLLVI3y6BK+P19GUsBV+sKj5ff83WwPjOoIuR k0NCwERi48knjDD2rP6vzCA2m4C6xI0bP8FsEQEziYOtf9i7GLk4mAWWMUnsPdHABuIIC8xk lHjx/w4rSBWLgKrEiU0TmUBsXqCOG/Oa2CCmykus3nAAbBInUPzf7t/sILaQQLJEy/rfLCCD JATus0k0/fkAdYakxMEVN1gmMPIuYGRYxSiUmVeWm5iZY6KXUZmXWaGXnJ+7iREY2Mtq/0Tv YPx0IfgQowAHoxIP74Mze9OFWBPLiitzDzFKcDArifByZu5JF+JNSaysSi3Kjy8qzUktPsQo zcGiJM5r9K08RUggPbEkNTs1tSC1CCbLxMEp1cBYO7Hbtv1PTnnl1/QjFxe3TSxfW2WhO43r 1I+i/wf3C6WJz9t2cMqP5UKlaVU2C6aZC/43SPCsv7fpwoa/dssC3rx/ai6k4P97s6VRj1bc gq0TjFa8fG67RGjyaSZdV+4XZcpq1W03Xbec+79sP8+Z5u0yni/jH3R9zXtjf7/aWLNn+yEv t+1nlFiKMxINtZiLihMBFJdnCmgCAAA= X-Brightmail-Tracker: H4sIAAAAAAAAA+NgFtrDLMWRmVeSWpSXmKPExsXC5WfdrLucaX+6Qf9jRYs569ewWXze8I/N 4uv6X8wWTz/1sVgcnnuS1eLyrjlsFvfW/Ge1OL9rLavFjqX7mCwuHVjAZHG89wCTxfx7n9ks Nm+aymxxfMpURovfP+awOfB7fG/tY/HYOesuu8eCTaUem1doeWxa1cnmsenTJHaPd+fOsXuc mPGbxeP9vqtsHotffGDy2PrLzqNx6jU2j8+b5AJ4o7hsUlJzMstSi/TtErgyXk9fxlLwxari 8/XXbA2M7wy6GDk5JARMJGb1f2UGsdkE1CVu3PgJZosImEkcbP3D3sXIxcEssIxJYu+JBjYQ R1hgJqPEi/93WEGqWARUJU5smsgEYvMCddyY18QGMVVeYvWGA2CTOIHi/3b/ZgexhQSSJVrW /2aZwMi1gJFhFaNIZl5ZbmJmjqlecXZGZV5mhV5yfu4mRmCYLqv9M3EH45fL7ocYBTgYlXh4 H5zZmy7EmlhWXJl7iFGCg1lJhJczc0+6EG9KYmVValF+fFFpTmrxIUZpDhYlcV6v8NQEIYH0 xJLU7NTUgtQimCwTB6dUA6Pi/8N/DF4/6VNO22JuWHxJQMp62mxfzyTWiXVzMvdemGr8WY7x W06zx4ubobOnCS+9+uOao6eg2p78a0f+O8/dWVdwRk+n9POX77qpzxiTb5vkrftso+UbnWLq 7qi3UDR2XrmweYRRMmfjnJeuPPOPOp0/JZMX1z9r+6zNvq9jCsNfuDPXLFJiKc5INNRiLipO BACYa2oITwIAAA== X-CFilter-Loop: Reflected X-Rspamd-Queue-Id: 00059180034 X-Stat-Signature: y1bbjhbw74yz74a3a1zj9wtowx16fwqs X-Rspam-User: X-Rspamd-Server: rspam10 X-HE-Tag: 1740571312-153931 X-HE-Meta: U2FsdGVkX1+O03uVno6XRSkDvHaMelXkaN3MYku1YgOjeIN19UFMt3SjdSQQyeDp64eO5x9GRHqSQslpsFm2Obelrx4AvresDcUIy2OEaHfWFAtHjNLb4HMOjpzUCUfXPTD2Tm2ZspPp/tKUcd5iO249p+hpRCVBOa+8USnd3Vm+nhtzZZGa2CovPffYZsNZ9B7Kk9yVJmykq09OXXCBOgyu9/cNqDbiff9bEtqWApy1wd1oSjT/cnLPgPgOXBPyPExS1ilWZnaSd9G3bDhytwwTV+jebeFSKSo9rcpNB2Km9bDPpPAuMbJH422HLowRgAKwDhM1pYOXWe/m5XdhApGIih0dA1W9GyLBLszVAp8M9PwSNA0eIUJ7o40mT4oHJI0hvkiPpzRhFrAuINnbqEcrkw7nwjMiDeSRrYrSuqczuJL5W1BV7OZlXJFq0fWdZH7X40vQQJmY/xFq78YRbZHDbq6KDdzjKQUqIlAIzaWkckHwHJaaZqYukUtCEmFUMnD930UXaX2253SyvTnbdV/MYXQTqwA1PNfll0kQS3fRqqjCOlJtU01WmwjkpiGHbKe/uBwHN2YoO0KoR/GCPCoT483LjJQ3lHU4ZjlpgEWtOgtkqG6odLh4ryI7ExYhv/w1/ivUhNEWZ8ItOEuKbpvbXQNRrzIypyoPsboZpokLrNEoM85/z3dhSEXJPraJgQ7w9jdB/N210zQJRNU+IWpKdGAorY4J1JaNp+PwZ86T+qmf4gTfqXk9hdPEPrsqOs9JTH1j8st4IrQ3Fn8BeTd6udoB5SZymVcCP1u0c4cP9ATMN9KDpCEPZ2ZY+iNjuQhmMj77UQFDJYr5h0UFgh8rjKi/I36ahUKUd8jX5BxT9jzcntdiaptWAGch6pVaHvq92iIUKn/WIt4pQXhHDyjKKG14HL9+kdbPV8mf7mx/3XLcXt5e1+Ey2dRw4IXFVNI9o8ly/xl4s2Hawb2 kRnEdu+8 6On6zxlhBSeY05Pu/3ZAlhSZHs92NyUXwYiiF9677a6cUIHIOnpQcVk6CgNzY6YGSSuiAnb5q4Ntp2WU= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: Do not perform tlb shootdown if the context is in preempt disable and there are already enough non luf pages, not to hurt preemptibility. Signed-off-by: Byungchul Park --- mm/compaction.c | 6 +++--- mm/internal.h | 5 +++-- mm/page_alloc.c | 27 +++++++++++++++------------ mm/page_isolation.c | 2 +- mm/page_reporting.c | 4 ++-- 5 files changed, 24 insertions(+), 20 deletions(-) diff --git a/mm/compaction.c b/mm/compaction.c index c87a1803b10e2..9098ddb04bbf5 100644 --- a/mm/compaction.c +++ b/mm/compaction.c @@ -606,7 +606,7 @@ static unsigned long isolate_freepages_block(struct compact_control *cc, page = pfn_to_page(blockpfn); - luf_takeoff_start(); + luf_takeoff_start(cc->zone); /* Isolate free pages. */ for (; blockpfn < end_pfn; blockpfn += stride, page += stride) { int isolated; @@ -1603,7 +1603,7 @@ static void fast_isolate_freepages(struct compact_control *cc) if (!area->nr_free) continue; - can_shootdown = luf_takeoff_start(); + can_shootdown = luf_takeoff_start(cc->zone); spin_lock_irqsave(&cc->zone->lock, flags); freelist = &area->free_list[MIGRATE_MOVABLE]; retry: @@ -2417,7 +2417,7 @@ static enum compact_result compact_finished(struct compact_control *cc) * luf_takeoff_{start,end}() is required to identify whether * this compaction context is tlb shootdownable for luf'd pages. */ - luf_takeoff_start(); + luf_takeoff_start(cc->zone); ret = __compact_finished(cc); luf_takeoff_end(cc->zone); diff --git a/mm/internal.h b/mm/internal.h index 53056ad7dade9..7c4198f5e22c3 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -1609,7 +1609,7 @@ static inline void accept_page(struct page *page) #endif /* CONFIG_UNACCEPTED_MEMORY */ #if defined(CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH) extern struct luf_batch luf_batch[]; -bool luf_takeoff_start(void); +bool luf_takeoff_start(struct zone *zone); void luf_takeoff_end(struct zone *zone); bool luf_takeoff_no_shootdown(void); bool luf_takeoff_check(struct zone *zone, struct page *page); @@ -1623,6 +1623,7 @@ static inline bool non_luf_pages_ok(struct zone *zone) return nr_free - nr_luf_pages > min_wm; } + unsigned short fold_unmap_luf(void); /* @@ -1709,7 +1710,7 @@ static inline bool can_luf_vma(struct vm_area_struct *vma) return true; } #else /* CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH */ -static inline bool luf_takeoff_start(void) { return false; } +static inline bool luf_takeoff_start(struct zone *zone) { return false; } static inline void luf_takeoff_end(struct zone *zone) {} static inline bool luf_takeoff_no_shootdown(void) { return true; } static inline bool luf_takeoff_check(struct zone *zone, struct page *page) { return true; } diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 0f986cfa4fe39..9a58d6f7a9609 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -623,22 +623,25 @@ compaction_capture(struct capture_control *capc, struct page *page, #endif /* CONFIG_COMPACTION */ #if defined(CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH) -static bool no_shootdown_context(void) +static bool no_shootdown_context(struct zone *zone) { /* - * If it performs with irq disabled, that might cause a deadlock. - * Avoid tlb shootdown in this case. + * Tries to avoid tlb shootdown if !preemptible(). However, it + * should be allowed under heavy memory pressure. */ + if (zone && non_luf_pages_ok(zone)) + return !(preemptible() && in_task()); + return !(!irqs_disabled() && in_task()); } /* * Can be called with zone lock released and irq enabled. */ -bool luf_takeoff_start(void) +bool luf_takeoff_start(struct zone *zone) { unsigned long flags; - bool no_shootdown = no_shootdown_context(); + bool no_shootdown = no_shootdown_context(zone); local_irq_save(flags); @@ -2669,7 +2672,7 @@ static bool unreserve_highatomic_pageblock(const struct alloc_context *ac, * luf_takeoff_{start,end}() is required for * get_page_from_free_area() to use luf_takeoff_check(). */ - luf_takeoff_start(); + luf_takeoff_start(zone); spin_lock_irqsave(&zone->lock, flags); for (order = 0; order < NR_PAGE_ORDERS; order++) { struct free_area *area = &(zone->free_area[order]); @@ -2874,7 +2877,7 @@ static int rmqueue_bulk(struct zone *zone, unsigned int order, unsigned long flags; int i; - luf_takeoff_start(); + luf_takeoff_start(zone); spin_lock_irqsave(&zone->lock, flags); for (i = 0; i < count; ++i) { struct page *page = __rmqueue(zone, order, migratetype, @@ -3500,7 +3503,7 @@ struct page *rmqueue_buddy(struct zone *preferred_zone, struct zone *zone, do { page = NULL; - luf_takeoff_start(); + luf_takeoff_start(zone); spin_lock_irqsave(&zone->lock, flags); if (alloc_flags & ALLOC_HIGHATOMIC) page = __rmqueue_smallest(zone, order, MIGRATE_HIGHATOMIC); @@ -3645,7 +3648,7 @@ static struct page *rmqueue_pcplist(struct zone *preferred_zone, struct page *page; unsigned long __maybe_unused UP_flags; - luf_takeoff_start(); + luf_takeoff_start(NULL); /* spin_trylock may fail due to a parallel drain or IRQ reentrancy. */ pcp_trylock_prepare(UP_flags); pcp = pcp_spin_trylock(zone->per_cpu_pageset); @@ -5268,7 +5271,7 @@ unsigned long alloc_pages_bulk_noprof(gfp_t gfp, int preferred_nid, if (unlikely(!zone)) goto failed; - luf_takeoff_start(); + luf_takeoff_start(NULL); /* spin_trylock may fail due to a parallel drain or IRQ reentrancy. */ pcp_trylock_prepare(UP_flags); pcp = pcp_spin_trylock(zone->per_cpu_pageset); @@ -7371,7 +7374,7 @@ unsigned long __offline_isolated_pages(unsigned long start_pfn, offline_mem_sections(pfn, end_pfn); zone = page_zone(pfn_to_page(pfn)); - luf_takeoff_start(); + luf_takeoff_start(zone); spin_lock_irqsave(&zone->lock, flags); while (pfn < end_pfn) { page = pfn_to_page(pfn); @@ -7489,7 +7492,7 @@ bool take_page_off_buddy(struct page *page) unsigned int order; bool ret = false; - luf_takeoff_start(); + luf_takeoff_start(zone); spin_lock_irqsave(&zone->lock, flags); for (order = 0; order < NR_PAGE_ORDERS; order++) { struct page *page_head = page - (pfn & ((1 << order) - 1)); diff --git a/mm/page_isolation.c b/mm/page_isolation.c index 521ed32bdbf67..70f938c0921ae 100644 --- a/mm/page_isolation.c +++ b/mm/page_isolation.c @@ -218,7 +218,7 @@ static void unset_migratetype_isolate(struct page *page, int migratetype) struct page *buddy; zone = page_zone(page); - luf_takeoff_start(); + luf_takeoff_start(zone); spin_lock_irqsave(&zone->lock, flags); if (!is_migrate_isolate_page(page)) goto out; diff --git a/mm/page_reporting.c b/mm/page_reporting.c index b23d3ed34ec07..83b66e7f0d257 100644 --- a/mm/page_reporting.c +++ b/mm/page_reporting.c @@ -170,7 +170,7 @@ page_reporting_cycle(struct page_reporting_dev_info *prdev, struct zone *zone, if (free_area_empty(area, mt)) return err; - can_shootdown = luf_takeoff_start(); + can_shootdown = luf_takeoff_start(zone); spin_lock_irq(&zone->lock); /* @@ -250,7 +250,7 @@ page_reporting_cycle(struct page_reporting_dev_info *prdev, struct zone *zone, /* update budget to reflect call to report function */ budget--; - luf_takeoff_start(); + luf_takeoff_start(zone); /* reacquire zone lock and resume processing */ spin_lock_irq(&zone->lock);