From patchwork Fri Apr 29 00:09:47 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Jiaqi Yan X-Patchwork-Id: 12831342 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 6B846C433F5 for ; Fri, 29 Apr 2022 00:09:55 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id BAA6B6B0074; Thu, 28 Apr 2022 20:09:54 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id B58D46B0075; Thu, 28 Apr 2022 20:09:54 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 9ACD86B0078; Thu, 28 Apr 2022 20:09:54 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (relay.a.hostedemail.com [64.99.140.24]) by kanga.kvack.org (Postfix) with ESMTP id 8BC406B0074 for ; Thu, 28 Apr 2022 20:09:54 -0400 (EDT) Received: from smtpin16.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay13.hostedemail.com (Postfix) with ESMTP id 5758E6058D for ; Fri, 29 Apr 2022 00:09:54 +0000 (UTC) X-FDA: 79407983508.16.D5F9DBA Received: from mail-pg1-f201.google.com (mail-pg1-f201.google.com [209.85.215.201]) by imf08.hostedemail.com (Postfix) with ESMTP id 22AD7160052 for ; Fri, 29 Apr 2022 00:09:46 +0000 (UTC) Received: by mail-pg1-f201.google.com with SMTP id t3-20020a656083000000b0039cf337edd6so3130871pgu.18 for ; Thu, 28 Apr 2022 17:09:53 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20210112; h=date:in-reply-to:message-id:mime-version:references:subject:from:to :cc; bh=Pmjf3nzz96Ndcqn7fnfXcV8alC2DVDHgj4fB9IqFuDw=; b=Zw7sa4jHLUwgCVdlMA3uTQiJfhMVrBRYJs7EUwRkG9DoC6sC9HBJ7j8KPd1EVk92+1 1tbUVMe8rJWJecLFXDVqSZrdLQwT2SX6lhgv5gr4W+/3dSnXx4lwXBcYsNgvHcBSkXvl PGV/gTd+ltsKQ35ZNMxVinEmjMlQLJk3Yz7qQ++8yTj9GfWU5et23Ix8oxffjen9DAEf CWZNAdgI0silkhZVr0JB/Kr5F8e+YY+saHKlS6YhbcVPj8WPYBEk8tLXS9y/0Mcc5QZP pBugjBXQZdCacEhW54seUT3ASBtu7ZPLUO0jSGQJJKblq6bMK+/iSACNoiFIHWc+6hhS j9uQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:date:in-reply-to:message-id:mime-version :references:subject:from:to:cc; bh=Pmjf3nzz96Ndcqn7fnfXcV8alC2DVDHgj4fB9IqFuDw=; b=5GuKr6JUgdb3o7W4aYufucHWce1kIU4Uju2LE3kFI63mv+eOLRKsgSz+P5eGLdxy47 d/GRUxVKjmqzJSPpaAnOHp9l1nusNhybtbZUo7y6k5weg3mdQK8QXDFKUv531cjWKl6z tkdIXdY5XSOrf5y8rxVry+7I0ZfpvcVrbwgvMNc+TXVYLOIb66y9opKLooWSalyxRP3x ZcPJzgQMwmmwDe7bDuCdnmkcgGlXS2Dq9ak6Ck74O2Yydb8R5bnPx/5D1FcS/Ojsdi39 W6x2tDljnw+WNpFgkX9A3lnszYlDSYCcWqg+2XzaKysipEkf7sozBl1V4vRm0sP8VvxP N/Cg== X-Gm-Message-State: AOAM533qvLUY38puw5lEuSj8QF1MHF7y6LGPKwI996H3ozhrFSv8SRTl YjZmughZMv57F735KSTL4l6/qA9jsDjNaQ== X-Google-Smtp-Source: ABdhPJwDuvSFmFg2lz46R146ifDrCpn7kV3u15eXXuT/1Owbza/2Lh99kU5Pds0Fd0tHWZyJAL1pXNgP7Ij2Pg== X-Received: from yjqkernel.c.googlers.com ([fda3:e722:ac3:cc00:7f:e700:c0a8:1837]) (user=jiaqiyan job=sendgmr) by 2002:a17:90b:3903:b0:1d2:c8ea:4c9c with SMTP id ob3-20020a17090b390300b001d2c8ea4c9cmr910200pjb.210.1651190992943; Thu, 28 Apr 2022 17:09:52 -0700 (PDT) Date: Thu, 28 Apr 2022 17:09:47 -0700 In-Reply-To: <20220429000947.2172219-1-jiaqiyan@google.com> Message-Id: <20220429000947.2172219-3-jiaqiyan@google.com> Mime-Version: 1.0 References: <20220429000947.2172219-1-jiaqiyan@google.com> X-Mailer: git-send-email 2.36.0.464.gb9c8b46e94-goog Subject: [RFC v2 2/2] mm: khugepaged: recover from poisoned file-backed memory From: Jiaqi Yan To: shy828301@gmail.com, tongtiangen@huawei.com Cc: linux-mm@kvack.org, tony.luck@intel.com, naoya.horiguchi@nec.com, kirill.shutemov@linux.intel.com, linmiaohe@huawei.com, juew@google.com, jiaqiyan@google.com X-Rspam-User: X-Rspamd-Server: rspam11 X-Rspamd-Queue-Id: 22AD7160052 X-Stat-Signature: 5wteke1p9mdgpu599ysouqbgqp1wq8oy Authentication-Results: imf08.hostedemail.com; dkim=pass header.d=google.com header.s=20210112 header.b=Zw7sa4jH; spf=pass (imf08.hostedemail.com: domain of 30CxrYggKCO8aZRhZpReXffXcV.TfdcZelo-ddbmRTb.fiX@flex--jiaqiyan.bounces.google.com designates 209.85.215.201 as permitted sender) smtp.mailfrom=30CxrYggKCO8aZRhZpReXffXcV.TfdcZelo-ddbmRTb.fiX@flex--jiaqiyan.bounces.google.com; dmarc=pass (policy=reject) header.from=google.com X-HE-Tag: 1651190986-396450 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: Make collapse_file roll back when copying pages failed. More concretely: * extract copying operations into a separate loop * postpone the updates for nr_none until both scanning and copying succeeded * postpone joining small xarray entries until both scanning and copying succeeded * postpone the update operations to NR_XXX_THPS * for non-SHMEM file, roll back filemap_nr_thps_inc if scan succeeded but copying failed Signed-off-by: Jiaqi Yan --- mm/khugepaged.c | 77 ++++++++++++++++++++++++++++++------------------- 1 file changed, 48 insertions(+), 29 deletions(-) diff --git a/mm/khugepaged.c b/mm/khugepaged.c index 8e69a0640e551..08f93f2ea9712 100644 --- a/mm/khugepaged.c +++ b/mm/khugepaged.c @@ -1710,7 +1710,7 @@ static void collapse_file(struct mm_struct *mm, { struct address_space *mapping = file->f_mapping; gfp_t gfp; - struct page *new_page; + struct page *new_page, *page, *tmp; pgoff_t index, end = start + HPAGE_PMD_NR; LIST_HEAD(pagelist); XA_STATE_ORDER(xas, &mapping->i_pages, start, HPAGE_PMD_ORDER); @@ -1766,7 +1766,7 @@ static void collapse_file(struct mm_struct *mm, xas_set(&xas, start); for (index = start; index < end; index++) { - struct page *page = xas_next(&xas); + page = xas_next(&xas); VM_BUG_ON(index != xas.xa_index); if (is_shmem) { @@ -1938,10 +1938,7 @@ static void collapse_file(struct mm_struct *mm, } nr = thp_nr_pages(new_page); - if (is_shmem) - __mod_lruvec_page_state(new_page, NR_SHMEM_THPS, nr); - else { - __mod_lruvec_page_state(new_page, NR_FILE_THPS, nr); + if (!is_shmem) { filemap_nr_thps_inc(mapping); /* * Paired with smp_mb() in do_dentry_open() to ensure @@ -1952,40 +1949,44 @@ static void collapse_file(struct mm_struct *mm, smp_mb(); if (inode_is_open_for_write(mapping->host)) { result = SCAN_FAIL; - __mod_lruvec_page_state(new_page, NR_FILE_THPS, -nr); filemap_nr_thps_dec(mapping); goto xa_locked; } } - if (nr_none) { - __mod_lruvec_page_state(new_page, NR_FILE_PAGES, nr_none); - if (is_shmem) - __mod_lruvec_page_state(new_page, NR_SHMEM, nr_none); - } - - /* Join all the small entries into a single multi-index entry */ - xas_set_order(&xas, start, HPAGE_PMD_ORDER); - xas_store(&xas, new_page); xa_locked: xas_unlock_irq(&xas); xa_unlocked: if (result == SCAN_SUCCEED) { - struct page *page, *tmp; - /* * Replacing old pages with new one has succeeded, now we - * need to copy the content and free the old pages. + * attempt to copy the contents. */ index = start; - list_for_each_entry_safe(page, tmp, &pagelist, lru) { + list_for_each_entry(page, &pagelist, lru) { while (index < page->index) { clear_highpage(new_page + (index % HPAGE_PMD_NR)); index++; } - copy_highpage(new_page + (page->index % HPAGE_PMD_NR), - page); + if (copy_highpage_mc(new_page + (page->index % HPAGE_PMD_NR), page)) { + result = SCAN_COPY_MC; + break; + } + index++; + } + while (result == SCAN_SUCCEED && index < end) { + clear_highpage(new_page + (page->index % HPAGE_PMD_NR)); + index++; + } + } + + if (result == SCAN_SUCCEED) { + /* + * Copying old pages to huge one has succeeded, now we + * need to free the old pages. + */ + list_for_each_entry_safe(page, tmp, &pagelist, lru) { list_del(&page->lru); page->mapping = NULL; page_ref_unfreeze(page, 1); @@ -1993,12 +1994,23 @@ static void collapse_file(struct mm_struct *mm, ClearPageUnevictable(page); unlock_page(page); put_page(page); - index++; } - while (index < end) { - clear_highpage(new_page + (index % HPAGE_PMD_NR)); - index++; + + xas_lock_irq(&xas); + if (is_shmem) + __mod_lruvec_page_state(new_page, NR_SHMEM_THPS, nr); + else + __mod_lruvec_page_state(new_page, NR_FILE_THPS, nr); + + if (nr_none) { + __mod_lruvec_page_state(new_page, NR_FILE_PAGES, nr_none); + if (is_shmem) + __mod_lruvec_page_state(new_page, NR_SHMEM, nr_none); } + /* Join all the small entries into a single multi-index entry. */ + xas_set_order(&xas, start, HPAGE_PMD_ORDER); + xas_store(&xas, new_page); + xas_unlock_irq(&xas); SetPageUptodate(new_page); page_ref_add(new_page, HPAGE_PMD_NR - 1); @@ -2014,9 +2026,9 @@ static void collapse_file(struct mm_struct *mm, khugepaged_pages_collapsed++; } else { - struct page *page; - - /* Something went wrong: roll back page cache changes */ + /* + * Something went wrong: roll back page cache changes + */ xas_lock_irq(&xas); mapping->nrpages -= nr_none; @@ -2049,6 +2061,13 @@ static void collapse_file(struct mm_struct *mm, xas_lock_irq(&xas); } VM_BUG_ON(nr_none); + /* + * Undo the updates of filemap_nr_thps_inc for non-SHMEM file only. + * This undo is not needed unless failure is due to SCAN_COPY_MC. + */ + if (!is_shmem && result == SCAN_COPY_MC) + filemap_nr_thps_dec(mapping); + xas_unlock_irq(&xas); new_page->mapping = NULL;