From patchwork Tue Apr 15 23:16:35 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Davidlohr Bueso X-Patchwork-Id: 14052927 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 57A76C369AB for ; Tue, 15 Apr 2025 23:17:12 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 1EC0D6B011F; Tue, 15 Apr 2025 19:16:53 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id E6DBE6B01FB; Tue, 15 Apr 2025 19:16:52 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id BFB0A6B01FC; Tue, 15 Apr 2025 19:16:52 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0011.hostedemail.com [216.40.44.11]) by kanga.kvack.org (Postfix) with ESMTP id 8E39A6B01F7 for ; Tue, 15 Apr 2025 19:16:52 -0400 (EDT) Received: from smtpin04.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay02.hostedemail.com (Postfix) with ESMTP id 73F08121A14 for ; Tue, 15 Apr 2025 23:16:53 +0000 (UTC) X-FDA: 83337840306.04.02118B9 Received: from giant.ash.relay.mailchannels.net (giant.ash.relay.mailchannels.net [23.83.222.68]) by imf07.hostedemail.com (Postfix) with ESMTP id 0C52340007 for ; Tue, 15 Apr 2025 23:16:50 +0000 (UTC) Authentication-Results: imf07.hostedemail.com; dkim=pass header.d=stgolabs.net header.s=dreamhost header.b=pdkeFYN+; arc=pass ("mailchannels.net:s=arc-2022:i=1"); spf=pass (imf07.hostedemail.com: domain of dave@stgolabs.net designates 23.83.222.68 as permitted sender) smtp.mailfrom=dave@stgolabs.net; dmarc=none ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1744759011; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=cZxBayk/19ttKico+tYOcz6z5i5bd/KpteFdnFTioS4=; b=Zv3cRjtjyrEAUNsswngexRyFuc2MbUJnDTjfeMGfK62rz462lAJTFXC780WaFlRf2ZiXsU tpb1blHRWBF0Im9TNOiHEO67aBzn5+x22A0pDKLItZSoZVpn7dffZmaaIlrdM0sFHQbuS4 f/Rd3ggUORj4sV7x2TJfyLwTTmLxdpk= ARC-Seal: i=2; s=arc-20220608; d=hostedemail.com; t=1744759011; a=rsa-sha256; cv=pass; b=Air2FIZ9tXSoey7QnwVWrjZK+PU5QpRaB1DFSPkRxWDIsBut9W+jeEUmsu9BQyyDJY56Pm FSg/3iX82XTrwU+o/ia4sWbfDaqpRoxckQrs4asxVgv6P8pJX69vPU8QsXbScxzSVyByN7 EYfXN52oGm3vkLnMxZRGfBteGJuX7cE= ARC-Authentication-Results: i=2; imf07.hostedemail.com; dkim=pass header.d=stgolabs.net header.s=dreamhost header.b=pdkeFYN+; arc=pass ("mailchannels.net:s=arc-2022:i=1"); spf=pass (imf07.hostedemail.com: domain of dave@stgolabs.net designates 23.83.222.68 as permitted sender) smtp.mailfrom=dave@stgolabs.net; dmarc=none X-Sender-Id: dreamhost|x-authsender|dave@stgolabs.net Received: from relay.mailchannels.net (localhost [127.0.0.1]) by relay.mailchannels.net (Postfix) with ESMTP id 8D86C844C70; Tue, 15 Apr 2025 23:16:49 +0000 (UTC) Received: from pdx1-sub0-mail-a273.dreamhost.com (trex-3.trex.outbound.svc.cluster.local [100.110.51.173]) (Authenticated sender: dreamhost) by relay.mailchannels.net (Postfix) with ESMTPA id 145DA844C93; Tue, 15 Apr 2025 23:16:49 +0000 (UTC) ARC-Seal: i=1; s=arc-2022; d=mailchannels.net; t=1744759009; a=rsa-sha256; cv=none; b=BhEJiJtFvkCwq8VAse19aBQiUxhkx3I+5R6HPsmqZf4e7aKwSiPPW986q2HE3kFNsRKNN+ CdfsNyJflEjavG33gCOm3aagchhCTRZDQsmsYDhQrVrS3YDc3D83CH8qXnfSqlJSWEFCwQ BAuiAebIKU36eG4L/2+jS4bFiSL6ap7/RY0lfPzATrkk+gG//gk2eCRvwHz/bvKAeUwy2O LztYFVMd88BfEaCPwhNC4l8tpMfN6HArgsnllkMyrgGuyaavWLatDR/edgfuhVMiictH0v HeefL/cKCyEyHtfkDDuAXLn2s6+YQU6DIBe1TIvPlvL9ViasagE8gE5uCYkpeA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=mailchannels.net; s=arc-2022; t=1744759009; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=cZxBayk/19ttKico+tYOcz6z5i5bd/KpteFdnFTioS4=; b=pwDyEFNKwZLvyNhomSgsA6Ch2JWc8J52qbKwVcFPj0Y6P7DTPH4++Hmj2H6BB/PL6KlNkN LzWPKmCveLupB2BZIPV6pfrWXrL1r3bW8ZJw43YHHLOtUCKKcGZLakiXg7C0B1fxMErIzW ifiBIjuHPfQIKiYdMcclddyU+UxOfndgfY4ZyGCOKUoIF9F+5GA7Y0Yv7uAeFGzt3tjPXl jCSz4lr6R8Y8Oqn6eNJOzJZXVa7GtBVpqV9vSxeCe6IotUVyiK/okDPirc1haqWpudz+F6 uqr1nVUCThwP4SDd+8+zflSksqc9IalFfqyCUMIN6j9NXJje2i7q2LNAV7x+0A== ARC-Authentication-Results: i=1; rspamd-5dd7f8b4cd-qrpj4; auth=pass smtp.auth=dreamhost smtp.mailfrom=dave@stgolabs.net X-Sender-Id: dreamhost|x-authsender|dave@stgolabs.net X-MC-Relay: Neutral X-MailChannels-SenderId: dreamhost|x-authsender|dave@stgolabs.net X-MailChannels-Auth-Id: dreamhost X-Abaft-Madly: 0f7b40494fe41933_1744759009477_58974177 X-MC-Loop-Signature: 1744759009477:425026726 X-MC-Ingress-Time: 1744759009477 Received: from pdx1-sub0-mail-a273.dreamhost.com (pop.dreamhost.com [64.90.62.162]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384) by 100.110.51.173 (trex/7.0.3); Tue, 15 Apr 2025 23:16:49 +0000 Received: from localhost.localdomain (ip72-199-50-187.sd.sd.cox.net [72.199.50.187]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits) server-digest SHA256) (No client certificate requested) (Authenticated sender: dave@stgolabs.net) by pdx1-sub0-mail-a273.dreamhost.com (Postfix) with ESMTPSA id 4Zcg5r0xWMz6g; Tue, 15 Apr 2025 16:16:48 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=stgolabs.net; s=dreamhost; t=1744759008; bh=cZxBayk/19ttKico+tYOcz6z5i5bd/KpteFdnFTioS4=; h=From:To:Cc:Subject:Date:Content-Transfer-Encoding; b=pdkeFYN+hvN6eTdJcNTCB0qKBgYHTg9M+wgnCHctw021Kq4TJBluwwznOkT4fiuae GioaEn1QRUFTyxxL9XY3Zj0XBm3PRV1Pm7rA+ilF/HBh6CP6lPiBPzOUkxXS1lA5iQ XCJgOHetlIP/kzdCjG5saON6RQClPMhB2BtsKBTwB6lsVAsqwhOq8297Sqc+kcQqwX VMNC4PQ2Eu+LJmiD//KZvW4/343ykgSTNa6QsEr0HI/YfOm5HitgLPdjITEbRGYsPE /aHt8eiK7FnTF0eOS6rGtlz2tQXeHNp0v3wVfnnolphExHswtWkYeQUEP+SG/puId5 60/AllAEwQfww== From: Davidlohr Bueso To: jack@suse.cz, tytso@mit.edu, adilger.kernel@dilger.ca, brauner@kernel.org Cc: mcgrof@kernel.org, willy@infradead.org, hare@suse.de, djwong@kernel.org, linux-ext4@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, Davidlohr Bueso , kernel test robot , syzbot+f3c6fda1297c748a7076@syzkaller.appspotmail.com Subject: [PATCH 7/7] mm/migrate: fix sleep in atomic for large folios and buffer heads Date: Tue, 15 Apr 2025 16:16:35 -0700 Message-Id: <20250415231635.83960-8-dave@stgolabs.net> X-Mailer: git-send-email 2.39.5 In-Reply-To: <20250415231635.83960-1-dave@stgolabs.net> References: <20250415231635.83960-1-dave@stgolabs.net> MIME-Version: 1.0 X-Stat-Signature: fw8tdtuekaga8smetfa6gcmyejzspgsj X-Rspam-User: X-Rspamd-Queue-Id: 0C52340007 X-Rspamd-Server: rspam08 X-HE-Tag: 1744759010-632994 X-HE-Meta: U2FsdGVkX18Yyg1FLz+F6gT0Rzp/BmhEiZJyftQNUeOGSXXTrPHQIENisOWHmnFrYZ43GALI+ZRoeaI3jqaP7GBZg2aFW6Rk/j2H6+NErkgCmgu4g67YF/m9ew4LW+a0i/3bOmXp3bovMInjySE/Gno6bO/K8flDXRlQbzQwebzf6XHByeMNBjlEDH5w8iDWtUht7xacrnVHyxHdZHi9a4Rkyv3t5P4g09VtYbMs9CwI8ez74kF2jCdleEQ760Z943yo8sJF6PfPyjCvU07p1dnM0xZDORgJ0qljMVW5NDun4SPGnG12shUY007EP2fAeNbriQcrneSWuRH3+m52XpAR4ngx7T45PzUK4oWLOpwS+ma/ISg6t7KxnW9fZqcg+D1d693DFF+wzBcAqAhQJ8EEK0V5tY9T+w6uHhojDKwYlooCXqfm/8CQes3sPSbVosKOzn0tRczHoz/LboRShqu5s2x7Lj+29K5jfdEn14zDG2nLHrb+c37AOcLJ7aAn0sEs+ZEgqVSbKBw0JlmX4U6t6DLSwvCg9fv99+8rMNEwPovvgnuXmE1lLKbH9qrOjxnrBS7bhAM35HjhCB9tYdEMVsMCj6vzwFtOhT1remzRu6zZaBNs2y7MX2MJ1TM0hV1PspYLYYwuzHY3tzKdXS4f92IzU9kvz5rx3EtHQYE0vmmKkbBbUQFpciSObdCPcG2UejrKWDLERaQh7KqC3Bqxq3bMcMkD74O+JJUjdNHv+0zGXeNwozArRpYV0/8U8viQpPqI6o0Om+/CLs5tL+9gunNHX4KVypCfjuAAjlUyECnRUdGTS0+YLhNr+nRNpzl2miidNw0mxzRhAOwsHZpXxAFrcdnW5B2JQBQwfO/O8qpF4FYCTRxHvAXTivFUIDoeZk21kKoqFZ8gAidV4w5IC6WBqMG+532nyM4SSbkcYvc7r30tB39xqiH3QS550EwsjJ5Ykc8rKvEbiSN 3dcqu5rh cNxtNfopt8mqSf/39T597scrK5qLLSkXkwGAZr+KBSYeqtlBxhAmYHMyNhK4Yeqdj+1FbD3JTiJqnOovwLab0lnuFFHFiyboLrPz6F5bSWGoEQylGhZwNnaM/Lj53zAe+96N0T9C5wThrnFDLEoRLM/k/KtzW7elPl51rgsCRxBt8sxZmVw72k3PtWRcoD7uAFV9fAV9uHTQwj62x5zo5ckqPd/vlK5Xzeu899wcPCrVIRr2tcWjY68u2VSv6AsBufpZYLl9fuAJBQHjcWa5l/5cwWvHmqLxLTc6qAze/aZyFgeNWj+N28DSzeHj+bADuejuVFHO+Hv0pMnKOk9payt/y1F1q/3SKerp0vOSfnl5InXM7/jh+bPnFM0p/V2a6mtSk2KhYSd+/0QHfe+6exUl1X4/eMLylECRL5FVZ2cIXlg6q0t3JLDyl6ERu9W3v9sAgLCAsR/L/oj3GpLqL21SYV5nuKl9CltpQjg5IUWaBULLO5eyGQLKSr4pjIAaeivtGMlL+SQZct0Gmu6fYzv1CajveO+yOxon+EFD7nOTTszE= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: The large folio + buffer head noref migration scenarios are being naughty and blocking while holding a spinlock. As a consequence of the pagecache lookup path taking the folio lock this serializes against migration paths, so they can wait for each other. For the private_lock atomic case, a new BH_Migrate flag is introduced which enables the lookup to bail. This allows the critical region of the private_lock on the migration path to be reduced to the way it was before ebdf4de5642fb6 ("mm: migrate: fix reference check race between __find_get_block() and migration"), that is covering the count checks. The scope is always noref migration. Reported-by: kernel test robot Reported-by: syzbot+f3c6fda1297c748a7076@syzkaller.appspotmail.com Closes: https://lore.kernel.org/oe-lkp/202503101536.27099c77-lkp@intel.com Fixes: 3c20917120ce61 ("block/bdev: enable large folio support for large logical block sizes") Co-developed-by: Luis Chamberlain Signed-off-by: Davidlohr Bueso Reviewed-by: Jan Kara --- fs/buffer.c | 12 +++++++++++- fs/ext4/ialloc.c | 3 ++- include/linux/buffer_head.h | 1 + mm/migrate.c | 8 +++++--- 4 files changed, 19 insertions(+), 5 deletions(-) 2.39.5 diff --git a/fs/buffer.c b/fs/buffer.c index f8e63885604b..b8e1e6e325cd 100644 --- a/fs/buffer.c +++ b/fs/buffer.c @@ -207,6 +207,15 @@ __find_get_block_slow(struct block_device *bdev, sector_t block, bool atomic) head = folio_buffers(folio); if (!head) goto out_unlock; + /* + * Upon a noref migration, the folio lock serializes here; + * otherwise bail. + */ + if (test_bit_acquire(BH_Migrate, &head->b_state)) { + WARN_ON(!atomic); + goto out_unlock; + } + bh = head; do { if (!buffer_mapped(bh)) @@ -1390,7 +1399,8 @@ lookup_bh_lru(struct block_device *bdev, sector_t block, unsigned size) /* * Perform a pagecache lookup for the matching buffer. If it's there, refresh * it in the LRU and mark it as accessed. If it is not present then return - * NULL + * NULL. Atomic context callers may also return NULL if the buffer is being + * migrated; similarly the page is not marked accessed either. */ static struct buffer_head * find_get_block_common(struct block_device *bdev, sector_t block, diff --git a/fs/ext4/ialloc.c b/fs/ext4/ialloc.c index 38bc8d74f4cc..e7ecc7c8a729 100644 --- a/fs/ext4/ialloc.c +++ b/fs/ext4/ialloc.c @@ -691,7 +691,8 @@ static int recently_deleted(struct super_block *sb, ext4_group_t group, int ino) if (!bh || !buffer_uptodate(bh)) /* * If the block is not in the buffer cache, then it - * must have been written out. + * must have been written out, or, most unlikely, is + * being migrated - false failure should be OK here. */ goto out; diff --git a/include/linux/buffer_head.h b/include/linux/buffer_head.h index c791aa9a08da..0029ff880e27 100644 --- a/include/linux/buffer_head.h +++ b/include/linux/buffer_head.h @@ -34,6 +34,7 @@ enum bh_state_bits { BH_Meta, /* Buffer contains metadata */ BH_Prio, /* Buffer should be submitted with REQ_PRIO */ BH_Defer_Completion, /* Defer AIO completion to workqueue */ + BH_Migrate, /* Buffer is being migrated (norefs) */ BH_PrivateStart,/* not a state bit, but the first bit available * for private allocation by other entities diff --git a/mm/migrate.c b/mm/migrate.c index 6e2488e5dbe4..c80591514e66 100644 --- a/mm/migrate.c +++ b/mm/migrate.c @@ -845,9 +845,11 @@ static int __buffer_migrate_folio(struct address_space *mapping, return -EAGAIN; if (check_refs) { - bool busy; + bool busy, migrating; bool invalidated = false; + migrating = test_and_set_bit_lock(BH_Migrate, &head->b_state); + VM_WARN_ON_ONCE(migrating); recheck_buffers: busy = false; spin_lock(&mapping->i_private_lock); @@ -859,12 +861,12 @@ static int __buffer_migrate_folio(struct address_space *mapping, } bh = bh->b_this_page; } while (bh != head); + spin_unlock(&mapping->i_private_lock); if (busy) { if (invalidated) { rc = -EAGAIN; goto unlock_buffers; } - spin_unlock(&mapping->i_private_lock); invalidate_bh_lrus(); invalidated = true; goto recheck_buffers; @@ -883,8 +885,7 @@ static int __buffer_migrate_folio(struct address_space *mapping, unlock_buffers: if (check_refs) - spin_unlock(&mapping->i_private_lock); + clear_bit_unlock(BH_Migrate, &head->b_state); bh = head; do { unlock_buffer(bh); --