From patchwork Sat Jun 15 03:49:54 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Baolin Wang X-Patchwork-Id: 13699223 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 0CD11C27C4F for ; Sat, 15 Jun 2024 03:55:23 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 540A66B0170; Fri, 14 Jun 2024 23:54:26 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 9DD976B02A9; Fri, 14 Jun 2024 23:54:25 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id C288F6B02B6; Fri, 14 Jun 2024 23:54:24 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0010.hostedemail.com [216.40.44.10]) by kanga.kvack.org (Postfix) with ESMTP id E68D76B02A8 for ; Fri, 14 Jun 2024 23:50:14 -0400 (EDT) Received: from smtpin15.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay03.hostedemail.com (Postfix) with ESMTP id 59AB7A0A59 for ; Sat, 15 Jun 2024 03:50:14 +0000 (UTC) X-FDA: 82231745148.15.683B650 Received: from out30-119.freemail.mail.aliyun.com (out30-119.freemail.mail.aliyun.com [115.124.30.119]) by imf05.hostedemail.com (Postfix) with ESMTP id 657A910000C for ; Sat, 15 Jun 2024 03:50:11 +0000 (UTC) Authentication-Results: imf05.hostedemail.com; dkim=pass header.d=linux.alibaba.com header.s=default header.b=tmhQLgk7; spf=pass (imf05.hostedemail.com: domain of baolin.wang@linux.alibaba.com designates 115.124.30.119 as permitted sender) smtp.mailfrom=baolin.wang@linux.alibaba.com; dmarc=pass (policy=none) header.from=linux.alibaba.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1718423410; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-transfer-encoding:content-transfer-encoding: in-reply-to:references:dkim-signature; bh=qyDcsqG1G9+ch9xsh4NPmqEryYrFj8x1o1BGt5MbPo0=; b=5dNF/pTrWIfgPlg3BlPK9opefOimQqFoGfvs+Zf/GcON1ij8syM7gSe2efNVLwyh5BZWv9 f81bNJgj+dlUsxw4m4Mw9ttH63orCr2fcVp/FDYVPZ0A8TtP2FvSVkWTUD/1as0ygDutfO rCz18twYzWaIAwQ9hD0XEROkDmcLBA8= ARC-Authentication-Results: i=1; imf05.hostedemail.com; dkim=pass header.d=linux.alibaba.com header.s=default header.b=tmhQLgk7; spf=pass (imf05.hostedemail.com: domain of baolin.wang@linux.alibaba.com designates 115.124.30.119 as permitted sender) smtp.mailfrom=baolin.wang@linux.alibaba.com; dmarc=pass (policy=none) header.from=linux.alibaba.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1718423410; a=rsa-sha256; cv=none; b=7/kUAaH3+XUd8ZVZKNWrB+FHjf5egEf75dKLBrkM5VkwvGZn7nG76iAR4XiYHiIVbAsEzN b/WQSvAr7BkYqiBhZFvAx017UHj/nL5tpLgfhoN0/+MeQbWWE0pWt9Oga8tU4C6IGG24Ml 1P12+XiGkCplOMzPOYMmRFDREFHj940= DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.alibaba.com; s=default; t=1718423408; h=From:To:Subject:Date:Message-Id:MIME-Version; bh=qyDcsqG1G9+ch9xsh4NPmqEryYrFj8x1o1BGt5MbPo0=; b=tmhQLgk7mCqb4wQbFPJ40TLp1ETTYTw5NeTkHEJp36bLZdgs6OBBOUPMHnK2+GJPfg4g4NWMGK95EFmjquFF3ujNyUH1GXbbcfcJbUrYrLDJvPbK7WTa+7KFIqGm8pQNecmn4+tW6gzrqJ5iNbWEp1Jm61q3bLVIOAFgdX7iwx4= X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R131e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=maildocker-contentspam033045075189;MF=baolin.wang@linux.alibaba.com;NM=1;PH=DS;RN=13;SR=0;TI=SMTPD_---0W8SZ6-6_1718423405; Received: from localhost(mailfrom:baolin.wang@linux.alibaba.com fp:SMTPD_---0W8SZ6-6_1718423405) by smtp.aliyun-inc.com; Sat, 15 Jun 2024 11:50:06 +0800 From: Baolin Wang To: akpm@linux-foundation.org, hughd@google.com Cc: willy@infradead.org, mhocko@kernel.org, roman.gushchin@linux.dev, shakeel.butt@linux.dev, muchun.song@linux.dev, hannes@cmpxchg.org, nphamcs@gmail.com, yosryahmed@google.com, baolin.wang@linux.alibaba.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH v3] mm: shmem: fix getting incorrect lruvec when replacing a shmem folio Date: Sat, 15 Jun 2024 11:49:54 +0800 Message-Id: X-Mailer: git-send-email 2.39.3 MIME-Version: 1.0 X-Stat-Signature: x9dmj999bqbrrhpawi6ckgw5tpige3bd X-Rspam-User: X-Rspamd-Queue-Id: 657A910000C X-Rspamd-Server: rspam02 X-HE-Tag: 1718423411-170482 X-HE-Meta: U2FsdGVkX1//hchWtOql5x4UN9xMAOBfM9nYFM5RuALQ68F67e1r3W0Wpig4aar3BUOwYap/OpuoPRRXlZ9AqDx6/AjYk+BsfPq2iWJ48VIPtgwO3PeJ5BGg/NIJOSCcXKBhjxAzjtkQztAA+d+FrSKXUoDYSp7c2LLsBSRMUqVXjkY7pqh2t0KSw0OGxajXLy0CKpj6OQkYyI/1ANeFzdWv/V2E8zyeDP5J8lwYRHSMjro2dKpOxZKL8Cgz+60FzqP6hUOFS9eiiLU7XYmn1Pcgp+2W0u51HTWKbtupkp62AiQNqruz4im/DY7tIiUhrKpEyM/PoIOe/+9XMfSuHHUuBYsxA/IjmrypBXFfJsTgcV/A6e3xiDD3VRvXyVNAzpeDN2fuH5/vHarIQt/iu9+Ap5NNOQDn3JMIM9/C5JENem8SgWs+NB0AMDEKoG9NI/5xUlSsV2lw4RswCB5UBm/NPvwS/Pnb3aH7kbpLOyJI7jrBBB3j7LaMlyUYwXEG9D3fa7NKKznD67X0leypH/qqsqeFXBn3Pyc1gwk6WULxNzLcXdzHEj4mHwbub4HuEY4dtLcNui8b6qXkX1Acbo2Tb2jKEwpiYFYsjtRc9OXty9oP90v61T+tcG3OJ6cQDzUOnIFvwntO66yzZWuqZMtIkUlzEOjXSPjBl4aRsM+wfdxjFuncueP9H2IgaX0KzmTlBCLoOQpAYoHAmUBj8COMJqW4akyXZdFCa747kmKnc+yF93v9NFceDQbEQzlgLVr4zrEF4LTrbU99u/L19tfUZ3TPErwAWtIyIkQdZDDwI+4zs/CuEXPr76c9l8LhK7m8oiQ250358dAGB5oxeziKCP0/9ECOJM4CiTG+pLXn7iN9Dp+2AWyFtxtyzwSYEzvhbucGZwkf0nUlvGVm+TMuWO/7IzSV4qcmXD64NxilXA4ofhZbD0QOFVaPrRg7Qa6YawEpMZiqtcIIoIC lqkYmUq6 JIC0DCXlxHvRLSO9z1q0baTKL2cUSQFpqDV/YUxGTGeqDGaGxSkI10r95oqcwHQGJvDucyVBmJ5T9Fs7HxYWSw6Fuce6fwfwnFifIol3SRYsQELL8tuoLLHcjOFVQ6S78UHsRZA2a0FIBgQFbKUzXEwc3bF2Q9EFK78STmBuew8rpMDdU+siTMVpBEAEp6tLSll3oJnqqA0ojliOq3H32qDDEdUV9F/V29YTnqohjK26kRdg= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: When testing shmem swapin, I encountered the warning below on my machine. The reason is that replacing an old shmem folio with a new one causes mem_cgroup_migrate() to clear the old folio's memcg data. As a result, the old folio cannot get the correct memcg's lruvec needed to remove itself from the LRU list when it is being freed. This could lead to possible serious problems, such as LRU list crashes due to holding the wrong LRU lock, and incorrect LRU statistics. To fix this issue, we can fallback to use the mem_cgroup_replace_folio() to replace the old shmem folio. [ 5241.100311] page: refcount:0 mapcount:0 mapping:0000000000000000 index:0x0 pfn:0x5d9960 [ 5241.100317] head: order:4 mapcount:0 entire_mapcount:0 nr_pages_mapped:0 pincount:0 [ 5241.100319] flags: 0x17fffe0000040068(uptodate|lru|head|swapbacked|node=0|zone=2|lastcpupid=0x3ffff) [ 5241.100323] raw: 17fffe0000040068 fffffdffd6687948 fffffdffd69ae008 0000000000000000 [ 5241.100325] raw: 0000000000000000 0000000000000000 00000000ffffffff 0000000000000000 [ 5241.100326] head: 17fffe0000040068 fffffdffd6687948 fffffdffd69ae008 0000000000000000 [ 5241.100327] head: 0000000000000000 0000000000000000 00000000ffffffff 0000000000000000 [ 5241.100328] head: 17fffe0000000204 fffffdffd6665801 ffffffffffffffff 0000000000000000 [ 5241.100329] head: 0000000a00000010 0000000000000000 00000000ffffffff 0000000000000000 [ 5241.100330] page dumped because: VM_WARN_ON_ONCE_FOLIO(!memcg && !mem_cgroup_disabled()) [ 5241.100338] ------------[ cut here ]------------ [ 5241.100339] WARNING: CPU: 19 PID: 78402 at include/linux/memcontrol.h:775 folio_lruvec_lock_irqsave+0x140/0x150 [...] [ 5241.100374] pc : folio_lruvec_lock_irqsave+0x140/0x150 [ 5241.100375] lr : folio_lruvec_lock_irqsave+0x138/0x150 [ 5241.100376] sp : ffff80008b38b930 [...] [ 5241.100398] Call trace: [ 5241.100399] folio_lruvec_lock_irqsave+0x140/0x150 [ 5241.100401] __page_cache_release+0x90/0x300 [ 5241.100404] __folio_put+0x50/0x108 [ 5241.100406] shmem_replace_folio+0x1b4/0x240 [ 5241.100409] shmem_swapin_folio+0x314/0x528 [ 5241.100411] shmem_get_folio_gfp+0x3b4/0x930 [ 5241.100412] shmem_fault+0x74/0x160 [ 5241.100414] __do_fault+0x40/0x218 [ 5241.100417] do_shared_fault+0x34/0x1b0 [ 5241.100419] do_fault+0x40/0x168 [ 5241.100420] handle_pte_fault+0x80/0x228 [ 5241.100422] __handle_mm_fault+0x1c4/0x440 [ 5241.100424] handle_mm_fault+0x60/0x1f0 [ 5241.100426] do_page_fault+0x120/0x488 [ 5241.100429] do_translation_fault+0x4c/0x68 [ 5241.100431] do_mem_abort+0x48/0xa0 [ 5241.100434] el0_da+0x38/0xc0 [ 5241.100436] el0t_64_sync_handler+0x68/0xc0 [ 5241.100437] el0t_64_sync+0x14c/0x150 [ 5241.100439] ---[ end trace 0000000000000000 ]--- Fixes: 85ce2c517ade ("memcontrol: only transfer the memcg data for migration") Signed-off-by: Baolin Wang Reviewed-by: Shakeel Butt --- Changes from v2: - Remove the less helpful comments, per Matthew. Changes from v1: - Add reviewed tag from Shakeel. - Update related comments, per Yosry. --- mm/memcontrol.c | 3 +-- mm/shmem.c | 2 +- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/mm/memcontrol.c b/mm/memcontrol.c index a811dfff10cd..c976db6a34c0 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -7780,8 +7780,7 @@ void __mem_cgroup_uncharge_folios(struct folio_batch *folios) * @new: Replacement folio. * * Charge @new as a replacement folio for @old. @old will - * be uncharged upon free. This is only used by the page cache - * (in replace_page_cache_folio()). + * be uncharged upon free. * * Both folios must be locked, @new->mapping must be set up. */ diff --git a/mm/shmem.c b/mm/shmem.c index 99bd3c34f0fb..4acaf02bfe44 100644 --- a/mm/shmem.c +++ b/mm/shmem.c @@ -1908,7 +1908,7 @@ static int shmem_replace_folio(struct folio **foliop, gfp_t gfp, xa_lock_irq(&swap_mapping->i_pages); error = shmem_replace_entry(swap_mapping, swap_index, old, new); if (!error) { - mem_cgroup_migrate(old, new); + mem_cgroup_replace_folio(old, new); __lruvec_stat_mod_folio(new, NR_FILE_PAGES, 1); __lruvec_stat_mod_folio(new, NR_SHMEM, 1); __lruvec_stat_mod_folio(old, NR_FILE_PAGES, -1);