From patchwork Mon Feb 24 16:55:47 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: David Hildenbrand X-Patchwork-Id: 13988498 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 6100AC021A4 for ; Mon, 24 Feb 2025 16:56:26 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 8F0766B0093; Mon, 24 Feb 2025 11:56:22 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 877A46B0095; Mon, 24 Feb 2025 11:56:22 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 6CA216B0096; Mon, 24 Feb 2025 11:56:22 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0012.hostedemail.com [216.40.44.12]) by kanga.kvack.org (Postfix) with ESMTP id 4BB3A6B0093 for ; Mon, 24 Feb 2025 11:56:22 -0500 (EST) Received: from smtpin09.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay10.hostedemail.com (Postfix) with ESMTP id 0389DC097B for ; Mon, 24 Feb 2025 16:56:21 +0000 (UTC) X-FDA: 83155441404.09.4B9D9F7 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) by imf29.hostedemail.com (Postfix) with ESMTP id BF7F8120012 for ; Mon, 24 Feb 2025 16:56:19 +0000 (UTC) Authentication-Results: imf29.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=gmzURY23; spf=pass (imf29.hostedemail.com: domain of dhildenb@redhat.com designates 170.10.129.124 as permitted sender) smtp.mailfrom=dhildenb@redhat.com; dmarc=pass (policy=none) header.from=redhat.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1740416179; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=v9Kk0sDLIOZEo9TGeBWuftLWUgUisx0QoodxZudqT8w=; b=atOhKYH4D5u2IpJ8uFiJ7S6wzqkCIqwvZjP3Si9NC3lTKdAV5+k9LOtETvHzXv52a98nRb BJaJlVlnzOcOix/aBDD6LxgHQk8cJMw5sqzUvyBziUZFrM6zvi4LBp9es/OG08TTU616lI WatvGkjWzt9W6KFUGJrLnd0JLZjxh1g= ARC-Authentication-Results: i=1; imf29.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=gmzURY23; spf=pass (imf29.hostedemail.com: domain of dhildenb@redhat.com designates 170.10.129.124 as permitted sender) smtp.mailfrom=dhildenb@redhat.com; dmarc=pass (policy=none) header.from=redhat.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1740416179; a=rsa-sha256; cv=none; b=kAqZ2mmp0rIkXzw+eKx/KkBqIMRkeH7/X1fc8N/as7gsUFYswA7u64whCuZbBPMcGAMF7J ZnVqhFdl5Xh4Yn+tvKJDoM5PP6dndnjI8WTylJARR7ROKdErECTjJ2sFNLYxqXx977YeWM GOzE6RIcOynH564uVgj8GTGFCdVzH5Y= DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1740416179; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=v9Kk0sDLIOZEo9TGeBWuftLWUgUisx0QoodxZudqT8w=; b=gmzURY23HMitCNJcpq5XOaXuPYE+zEo6bUmbGTUszPCwe3Q4ZZ5G5pCZbT72RJJtrg/ZhT YU3N8e+odeHfKo3gtdDd47OViXmkSg3eWGTQpFF7eTgJMeJD7mulZEVckwwXL1M1nl3cyC Oz3hdYfl8BcF+ctOvFeo9HKIjKKHc6M= Received: from mail-wm1-f69.google.com (mail-wm1-f69.google.com [209.85.128.69]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_256_GCM_SHA384) id us-mta-463-xGRp438EOx65OZxnqyRDow-1; Mon, 24 Feb 2025 11:56:17 -0500 X-MC-Unique: xGRp438EOx65OZxnqyRDow-1 X-Mimecast-MFC-AGG-ID: xGRp438EOx65OZxnqyRDow_1740416176 Received: by mail-wm1-f69.google.com with SMTP id 5b1f17b1804b1-4388eee7073so24508825e9.0 for ; Mon, 24 Feb 2025 08:56:17 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1740416176; x=1741020976; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=v9Kk0sDLIOZEo9TGeBWuftLWUgUisx0QoodxZudqT8w=; b=rxGi3tNPTBP5I2P6gjtVJZXhsFvdsh3IBCKubJHpL1nYb9V/25fD8/qrQ2hToGkNxc j326n1S35mDruL3CZJP52r0+zd2btFUbHy16gBxirCjp0DfUKci0a4NkoeIgQMmdXR3C tXaG51Ng4Fj2mh8oxz2Dqq1yJR9NhsU8Vjnulu9OBT6qfgaSzaXokoBN0ea3y6jLZ35Y lijL61/b4s2kmqKuThQhbGpM1Tn+z8EaUIpRH5Fp5fu4DkDi8EEABsqHrpitFZtTtgmz 7yCEm6Uigg7rmbr4oYjvzTm0B34HtsaXxSpeewp/iAHbPEjtcedgBh5XS7yTqhpl6Crb uaow== X-Forwarded-Encrypted: i=1; AJvYcCUbHJOugdsdsKkDNfms4xFXOiZjP2Np0JySVsFfdRHFr4hyhzQaEMvmUJdNUG5A2jJ1E6DUPevdjg==@kvack.org X-Gm-Message-State: AOJu0Yw2k2pLJDgLKVeiXarA41xI/qZ0pf6yPyDymTTzgc8YBkLfxGIc cx1WwDKr4q4/K7kgkDaqHhWU9mS3Jz6k1lDLLGQyGIbSG4rCtQPEcccjvBSqIVPsSF1Q6aArcUd srrcZpVjTc4a30+G+rYl7g1UOKgLMMuYdH8yL2EpIOZ7Ghp9e X-Gm-Gg: ASbGncsSKtxbl1fF3YI8392uGoWM7b3mA34CtOlAk0iEwZeUZurlpPsq0NivAjpG5f1 1IgqVjH2jqqtj9h56PVFHErdvDIIKnG6BbxiLUiwajAyRlBC6x/kuPIGCoQyjCCfTpPWHdFT4DF 9BF8eDZF9IMZX1QZwTynCIXSMWYtziWT4GxGA/oCknahCeY7RD23+qNRy4QmQTYGsiaefEbZv47 OnzEdhZVsvWW4nkCmdc8L4q8ryg6s8PdwzAG8Sbqb9iW55wcSWFTF5vZTXOZQL0USacJM2Prp53 GYU6SjTYYdasG6WYr981QQE2tZxf3D2yuu1YU7XtlA== X-Received: by 2002:a05:600c:5949:b0:439:8345:17a6 with SMTP id 5b1f17b1804b1-439a30d38e9mr130585495e9.12.1740416176144; Mon, 24 Feb 2025 08:56:16 -0800 (PST) X-Google-Smtp-Source: AGHT+IGy0+PSAOTGdTFyYGF1ZzYS11kOEi6LhvuqHcevrUwSSRUG4LJeLXU4Rq4oj/c975sX5HkQ4A== X-Received: by 2002:a05:600c:5949:b0:439:8345:17a6 with SMTP id 5b1f17b1804b1-439a30d38e9mr130585275e9.12.1740416175731; Mon, 24 Feb 2025 08:56:15 -0800 (PST) Received: from localhost (p4ff234b6.dip0.t-ipconnect.de. [79.242.52.182]) by smtp.gmail.com with UTF8SMTPSA id 5b1f17b1804b1-439b02ce404sm113163905e9.7.2025.02.24.08.56.14 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Mon, 24 Feb 2025 08:56:15 -0800 (PST) From: David Hildenbrand To: linux-kernel@vger.kernel.org Cc: linux-doc@vger.kernel.org, cgroups@vger.kernel.org, linux-mm@kvack.org, linux-fsdevel@vger.kernel.org, linux-api@vger.kernel.org, David Hildenbrand , Andrew Morton , "Matthew Wilcox (Oracle)" , Tejun Heo , Zefan Li , Johannes Weiner , =?utf-8?q?Michal_Koutn=C3=BD?= , Jonathan Corbet , Andy Lutomirski , Thomas Gleixner , Ingo Molnar , Borislav Petkov , Dave Hansen , Muchun Song , "Liam R. Howlett" , Lorenzo Stoakes , Vlastimil Babka , Jann Horn Subject: [PATCH v2 05/20] mm: move _pincount in folio to page[2] on 32bit Date: Mon, 24 Feb 2025 17:55:47 +0100 Message-ID: <20250224165603.1434404-6-david@redhat.com> X-Mailer: git-send-email 2.48.1 In-Reply-To: <20250224165603.1434404-1-david@redhat.com> References: <20250224165603.1434404-1-david@redhat.com> MIME-Version: 1.0 X-Mimecast-Spam-Score: 0 X-Mimecast-MFC-PROC-ID: j8FX-8JVjs0GPzZVRBJkoBNQvT7R2AzhcvDuBFq1ccg_1740416176 X-Mimecast-Originator: redhat.com content-type: text/plain; charset="US-ASCII"; x-default=true X-Rspamd-Queue-Id: BF7F8120012 X-Stat-Signature: g8dawwfes99epn43u99o6mgk56e6ibu6 X-Rspam-User: X-Rspamd-Server: rspam01 X-HE-Tag: 1740416179-122205 X-HE-Meta: U2FsdGVkX1/m+nWOlCIThvYJ77kd4I8aaQXiJJfqgtFzyMeamKJHZC4hodiAtJu250R34Qw/UAfiltYDU32Irh6ktsfDDUtCJtDhx2L6MeVSdNEKWTnvjv5dEBoCwY1qXVQBk/0zQP9Q5ZnjRCDZsSCN7G+XgIru8SvPv2EO4Nr3c+OTVcDE/ZsO2b9Zi373e/wRim8wh+Id8+D/Nw/1E0Amnnf2vEUYZ7Ok266mN/F+TKYL96QS13l4HuqmsUT5rfLhp7GyMTKrdFWn0TmJrqFP2QG2quiEB6/C3Jf1/uwWUIBqCGXZM6d7sILKxD0DOS7EDbc1oL52Vd17p5pMH5SO8EJgOm+XEZ0hMzDIGXyJ3SqKoA9PQVD3dzg0Cwt02n+pISmUpDwKmHUqDsJMZocUdadzMf1QjAp332UCuKx7IWHGE8ntmxZKmp+ISBNkUkgeVMFiJ2lHcTcmMt3f1EDPT7Qhw5LCabaHm4HdimYg5F+p0+8zp+uMJHxwgJ5Dtv/3wsFemq0P24uOR41Wxmo7wnY1Yl/6zt/Wg2gqkSHrpvm4cjv1qa+OjJwdHFQoJbJIQpiTnMMnMumUgPHW3teVkMBi1RJHQl4XF7imS70MxR60RlKXHjOAMVGnlw+G5cKpTKnbsAxdW0XvmCOmuhnpc2AAB887FqiTXWkRflWFOUzxeKpTSONgr5cLM2eyLdXyigEVJmei/0p3nQXIDtT7Kzh1JzUp8+O2J7xluq3ZxmVIy3baJF15VupZH5EPGH0+Z6yRXaGRhXBWtG6Kk69jZJ0vwp8l+R8nu3P4qQ+YxEsSMhQKqL/G/1JD4cqX3HvOVIrAxdOZpfVGFP5o1JMCeRgchZh/3opa4zmrYy3BWRyo3Rxt+7K4w4ob5VnAe3i2ZIPxPafiL3T4Z4D7BmjwwdmAvypRnbYxiMtS0fPZuvcuz34QoyLtuP8PrdAXK7OUxXOSDAgGeHyZza3 uwZUBlgi urD3qQwCGUP2nJcHJYhW9oTxJBuuRDCBR43jwtJDrlorAPbH+AETXjJnpuoyJcAVFmiCD2dFSyZEcZv/igDaLPatMLMCb7xOk6dJ5eWMhx38IUfprWBfoKBUJUiT5wWamUYKy1Osmm3a8jm9JIX+rIy1y1nRBq4gC/u+ht2Soy5eQXGpybqonQtomBgn78zvGjWExM5QTfFpCCjxp3PwOqikY90c2CxWxyFaywVtOmSVX2dL1u6CNFmMO26u8DBX4p1MkZDN5Q/yVeG0Ut9f1ruCLsNEivhW/Rm76bAgG/B23gaanjT0gjOd1lg+n3tnHDDQf0h/XU5kOu9dTAhHMMnAWekEEy2ln4GCsDXRf0jdzfGUhKB6EQ0q1Qf+M3mrSJGTtLHxXfXFdsodaFuqHgn5MNa5RzDjeVtxVs92bykZgJEkzGo3xgKabbJYTpbNW22jr93pL6R2bpntZY2oa3UhVNeRvy+FrAgDZ6e/DyFL/IuSSNydMjeMhsk8wAaSDXkVGRLinTgo6xUOAyOFfv6AShV0VFruYaZvvMh55TpXCQslrgSxUPjXZwe+sF7m5fPETKf/vjGHKbGs= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: Let's free up some space on 32bit in page[1] by moving the _pincount to page[2]. For order-1 folios (never anon folios!) on 32bit, we will now also use the GUP_PIN_COUNTING_BIAS approach. A fully-mapped order-1 folio requires 2 references. With GUP_PIN_COUNTING_BIAS being 1024, we'd detect such folios as "maybe pinned" with 512 full mappings, instead of 1024 for order-0. As anon folios are out of the picture (which are the most relevant users of checking for pinnings on *mapped* pages) and we are talking about 32bit, this is not expected to cause any trouble. In __dump_page(), copy one additional folio page if we detect a folio with an order > 1, so we can dump the pincount on order > 1 folios reliably. Note that THPs on 32bit are not particularly common (and we don't care too much about performance), but we want to keep it working reliably, because likely we want to use large folios there as well in the future, independent of PMD leaf support. Once we dynamically allocate "struct folio", fortunately the 32bit specifics will likely go away again; even small folios could then have a pincount and folio_has_pincount() would essentially always return "true". Signed-off-by: David Hildenbrand --- include/linux/mm.h | 11 +++++++++-- include/linux/mm_types.h | 5 +++++ mm/debug.c | 10 +++++++++- mm/gup.c | 8 ++++---- mm/internal.h | 3 ++- mm/page_alloc.c | 14 +++++++++++--- 6 files changed, 40 insertions(+), 11 deletions(-) diff --git a/include/linux/mm.h b/include/linux/mm.h index f6b6373a864dd..1a4ee028a851e 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -2004,6 +2004,13 @@ static inline struct folio *pfn_folio(unsigned long pfn) return page_folio(pfn_to_page(pfn)); } +static inline bool folio_has_pincount(const struct folio *folio) +{ + if (IS_ENABLED(CONFIG_64BIT)) + return folio_test_large(folio); + return folio_order(folio) > 1; +} + /** * folio_maybe_dma_pinned - Report if a folio may be pinned for DMA. * @folio: The folio. @@ -2020,7 +2027,7 @@ static inline struct folio *pfn_folio(unsigned long pfn) * get that many refcounts, and b) all the callers of this routine are * expected to be able to deal gracefully with a false positive. * - * For large folios, the result will be exactly correct. That's because + * For most large folios, the result will be exactly correct. That's because * we have more tracking data available: the _pincount field is used * instead of the GUP_PIN_COUNTING_BIAS scheme. * @@ -2031,7 +2038,7 @@ static inline struct folio *pfn_folio(unsigned long pfn) */ static inline bool folio_maybe_dma_pinned(struct folio *folio) { - if (folio_test_large(folio)) + if (folio_has_pincount(folio)) return atomic_read(&folio->_pincount) > 0; /* diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h index 1d9c68c551d42..31f466d8485bc 100644 --- a/include/linux/mm_types.h +++ b/include/linux/mm_types.h @@ -387,7 +387,9 @@ struct folio { atomic_t _large_mapcount; atomic_t _entire_mapcount; atomic_t _nr_pages_mapped; +#ifdef CONFIG_64BIT atomic_t _pincount; +#endif /* CONFIG_64BIT */ }; unsigned long _usable_1[4]; }; @@ -406,6 +408,9 @@ struct folio { unsigned long _head_2; /* public: */ struct list_head _deferred_list; +#ifndef CONFIG_64BIT + atomic_t _pincount; +#endif /* !CONFIG_64BIT */ /* private: the union with struct page is transitional */ }; struct page __page_2; diff --git a/mm/debug.c b/mm/debug.c index 2d1bd67d957bc..83ef3bd0ccd32 100644 --- a/mm/debug.c +++ b/mm/debug.c @@ -79,12 +79,17 @@ static void __dump_folio(struct folio *folio, struct page *page, folio_ref_count(folio), mapcount, mapping, folio->index + idx, pfn); if (folio_test_large(folio)) { + int pincount = 0; + + if (folio_has_pincount(folio)) + pincount = atomic_read(&folio->_pincount); + pr_warn("head: order:%u mapcount:%d entire_mapcount:%d nr_pages_mapped:%d pincount:%d\n", folio_order(folio), folio_mapcount(folio), folio_entire_mapcount(folio), folio_nr_pages_mapped(folio), - atomic_read(&folio->_pincount)); + pincount); } #ifdef CONFIG_MEMCG @@ -146,6 +151,9 @@ static void __dump_page(const struct page *page) if (idx < MAX_FOLIO_NR_PAGES) { memcpy(&folio, foliop, 2 * sizeof(struct page)); nr_pages = folio_nr_pages(&folio); + if (nr_pages > 1) + memcpy(&folio.__page_2, &foliop->__page_2, + sizeof(struct page)); foliop = &folio; } diff --git a/mm/gup.c b/mm/gup.c index e5040657870ea..2944fe8cf3174 100644 --- a/mm/gup.c +++ b/mm/gup.c @@ -109,7 +109,7 @@ static void gup_put_folio(struct folio *folio, int refs, unsigned int flags) if (is_zero_folio(folio)) return; node_stat_mod_folio(folio, NR_FOLL_PIN_RELEASED, refs); - if (folio_test_large(folio)) + if (folio_has_pincount(folio)) atomic_sub(refs, &folio->_pincount); else refs *= GUP_PIN_COUNTING_BIAS; @@ -164,7 +164,7 @@ int __must_check try_grab_folio(struct folio *folio, int refs, * Increment the normal page refcount field at least once, * so that the page really is pinned. */ - if (folio_test_large(folio)) { + if (folio_has_pincount(folio)) { folio_ref_add(folio, refs); atomic_add(refs, &folio->_pincount); } else { @@ -223,7 +223,7 @@ void folio_add_pin(struct folio *folio) * page refcount field at least once, so that the page really is * pinned. */ - if (folio_test_large(folio)) { + if (folio_has_pincount(folio)) { WARN_ON_ONCE(atomic_read(&folio->_pincount) < 1); folio_ref_inc(folio); atomic_inc(&folio->_pincount); @@ -575,7 +575,7 @@ static struct folio *try_grab_folio_fast(struct page *page, int refs, * is pinned. That's why the refcount from the earlier * try_get_folio() is left intact. */ - if (folio_test_large(folio)) + if (folio_has_pincount(folio)) atomic_add(refs, &folio->_pincount); else folio_ref_add(folio, diff --git a/mm/internal.h b/mm/internal.h index 7f6d5def00fa0..d33db24c8b17b 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -723,7 +723,8 @@ static inline void prep_compound_head(struct page *page, unsigned int order) atomic_set(&folio->_large_mapcount, -1); atomic_set(&folio->_entire_mapcount, -1); atomic_set(&folio->_nr_pages_mapped, 0); - atomic_set(&folio->_pincount, 0); + if (IS_ENABLED(CONFIG_64BIT) || order > 1) + atomic_set(&folio->_pincount, 0); if (order > 1) INIT_LIST_HEAD(&folio->_deferred_list); } diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 05a2a9492cdb0..3dff99cc54161 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -959,9 +959,11 @@ static int free_tail_page_prepare(struct page *head_page, struct page *page) bad_page(page, "nonzero nr_pages_mapped"); goto out; } - if (unlikely(atomic_read(&folio->_pincount))) { - bad_page(page, "nonzero pincount"); - goto out; + if (IS_ENABLED(CONFIG_64BIT)) { + if (unlikely(atomic_read(&folio->_pincount))) { + bad_page(page, "nonzero pincount"); + goto out; + } } break; case 2: @@ -970,6 +972,12 @@ static int free_tail_page_prepare(struct page *head_page, struct page *page) bad_page(page, "on deferred list"); goto out; } + if (!IS_ENABLED(CONFIG_64BIT)) { + if (unlikely(atomic_read(&folio->_pincount))) { + bad_page(page, "nonzero pincount"); + goto out; + } + } break; case 3: /* the third tail page: hugetlb specifics overlap ->mappings */