From patchwork Thu Dec 7 19:24:06 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Nhat Pham X-Patchwork-Id: 13483868 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id A3B29C10DC3 for ; Thu, 7 Dec 2023 19:24:11 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 12F7F6B0081; Thu, 7 Dec 2023 14:24:11 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 0DFB46B0083; Thu, 7 Dec 2023 14:24:11 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id EE9626B0085; Thu, 7 Dec 2023 14:24:10 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0011.hostedemail.com [216.40.44.11]) by kanga.kvack.org (Postfix) with ESMTP id DEA076B0082 for ; Thu, 7 Dec 2023 14:24:10 -0500 (EST) Received: from smtpin03.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay08.hostedemail.com (Postfix) with ESMTP id 9803E14020C for ; Thu, 7 Dec 2023 19:24:10 +0000 (UTC) X-FDA: 81540997860.03.32A9D1E Received: from mail-pl1-f175.google.com (mail-pl1-f175.google.com [209.85.214.175]) by imf29.hostedemail.com (Postfix) with ESMTP id CAAE6120020 for ; Thu, 7 Dec 2023 19:24:08 +0000 (UTC) Authentication-Results: imf29.hostedemail.com; dkim=pass header.d=gmail.com header.s=20230601 header.b=TnR554p4; spf=pass (imf29.hostedemail.com: domain of nphamcs@gmail.com designates 209.85.214.175 as permitted sender) smtp.mailfrom=nphamcs@gmail.com; dmarc=pass (policy=none) header.from=gmail.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1701977048; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-transfer-encoding:content-transfer-encoding: in-reply-to:references:dkim-signature; bh=nVsNmgmDnzQDG1/LGV7hJiMxT85AaGwagKzGkmoxER0=; b=ndYnsYFU3Uqjm6M/2gV9zazclJ/FQLVfRtAbBNF3It/QofIJg5WOKrsAtjkQtzBTwUUX/N M2ZNqb+JxZ9boQfSDTaBCul0CE+u2d0s/EmSNRM3nj5qUTRgIaVfm/XRvZYTLrX3CHaDf3 6ozAYp9lTJSTA+a3dwv1PYpn5WTO6nU= ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1701977048; a=rsa-sha256; cv=none; b=anagJPie+YZS7CZOHEACVFYJICTgmBzyfLw8vMZF28DBsKmoLAKs8QdSIkaeX0R0mCcZgc ujhQzfoD09ymdGvwgzqeSywACCHxWEhz+Gac/av2Ssegf6dplPmSoLWGlub5V23n+g/kre NRjv3GoYYwElAxHMR1s6TLrxMuHzVFU= ARC-Authentication-Results: i=1; imf29.hostedemail.com; dkim=pass header.d=gmail.com header.s=20230601 header.b=TnR554p4; spf=pass (imf29.hostedemail.com: domain of nphamcs@gmail.com designates 209.85.214.175 as permitted sender) smtp.mailfrom=nphamcs@gmail.com; dmarc=pass (policy=none) header.from=gmail.com Received: by mail-pl1-f175.google.com with SMTP id d9443c01a7336-1d098b87eeeso11512125ad.0 for ; Thu, 07 Dec 2023 11:24:08 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20230601; t=1701977047; x=1702581847; darn=kvack.org; h=content-transfer-encoding:mime-version:message-id:date:subject:cc :to:from:from:to:cc:subject:date:message-id:reply-to; bh=nVsNmgmDnzQDG1/LGV7hJiMxT85AaGwagKzGkmoxER0=; b=TnR554p41DTGU9ArU64ssnodNqNi9p3gAd9dRP+phzlkzN3BPqWD3TOP3lE4UeUrb3 r7Oj9gQlV7Orc1H1UHYKR2aolgtHtFqahwft+Uel6NTa7gaSFuyWNbtT+4ZrtBq53Fw8 dDOcCOw6Mf5FrONPTVRuS2s+30jOvXZh1BP5ocoYn2gXWSFJoFrUMcHL9/zEdoO2LPxd /ShYwbEbPgvJVOPcf1qyxiH3/DjeTSk0IX7MqSyTAH6XrLiypIwAMAihvA22WEVoTKO/ mmz3YtOnc/8wlhN/oGYPSG3SFOqD+WBF4yqUTcNFqrUxTaD6XWX6EVWWhcIztrv3TigT mrDg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1701977047; x=1702581847; h=content-transfer-encoding:mime-version:message-id:date:subject:cc :to:from:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=nVsNmgmDnzQDG1/LGV7hJiMxT85AaGwagKzGkmoxER0=; b=eA+4qFgXEE6b79/Doo0ILQP+I18vFNNnteUHE3qb8+TiIAB0B43t5Lorv5mMvEGSBn Me7NSnClZzKppUa1sgSZePp9Q5Yvte4cwAOAajEC8FODiLHIr+8JpalS9cCaZwmUCDeB lclDmPd6/ICyUNvUcUy4pMt1NSPQMyFLhpFhZt+Kxs2+Ifhrhbf7Vi9/AhlNWQKw4GkV NX60Va0Xa+xr04AJYfBkhyJwLgaYbqj32NO0bZWj6bcUoxoxuekYDO4r2YER6wt8Hd3J UZn4WJMMhIBMDTOgh2dTxN+ZOW2IyqUu87hwvxn1doV//xfm2aqmzdebDixP9Eg0pPXW X1Cw== X-Gm-Message-State: AOJu0Yw1j3GnxhtamZ0C6hnfOIYFGTImIce4/5IA4tfPYmMEqrRrzC2S FJQdln7TYepDTf5TiBpBuZk= X-Google-Smtp-Source: AGHT+IGUaegnoMWjY0e1XMRMVTrS0hgGoykO4U6+h1XcID6tzuveB11DDmljtRl89bro9Ih1KOiCfw== X-Received: by 2002:a17:902:d2cd:b0:1d0:6ffd:9e22 with SMTP id n13-20020a170902d2cd00b001d06ffd9e22mr3389722plc.116.1701977047355; Thu, 07 Dec 2023 11:24:07 -0800 (PST) Received: from localhost (fwdproxy-prn-112.fbsv.net. [2a03:2880:ff:70::face:b00c]) by smtp.gmail.com with ESMTPSA id b8-20020a170903228800b001d077c9a9acsm164633plh.185.2023.12.07.11.24.06 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 07 Dec 2023 11:24:07 -0800 (PST) From: Nhat Pham To: akpm@linux-foundation.org Cc: tj@kernel.org, lizefan.x@bytedance.com, hannes@cmpxchg.org, cerasuolodomenico@gmail.com, yosryahmed@google.com, sjenning@redhat.com, ddstreet@ieee.org, vitaly.wool@konsulko.com, mhocko@kernel.org, roman.gushchin@linux.dev, shakeelb@google.com, muchun.song@linux.dev, hughd@google.com, corbet@lwn.net, konrad.wilk@oracle.com, senozhatsky@chromium.org, rppt@kernel.org, linux-mm@kvack.org, kernel-team@meta.com, linux-kernel@vger.kernel.org, linux-doc@vger.kernel.org, david@ixit.cz, chrisl@kernel.org Subject: [PATCH v6] zswap: memcontrol: implement zswap writeback disabling Date: Thu, 7 Dec 2023 11:24:06 -0800 Message-Id: <20231207192406.3809579-1-nphamcs@gmail.com> X-Mailer: git-send-email 2.34.1 MIME-Version: 1.0 X-Stat-Signature: wrbzc3wxuj6nziba6aj6na1uyg97qh5h X-Rspamd-Server: rspam10 X-Rspamd-Queue-Id: CAAE6120020 X-Rspam-User: X-HE-Tag: 1701977048-65660 X-HE-Meta: U2FsdGVkX19R2/M0MFLgzq2eMFyiow7Mh95/bJ1Jli78dp6FMAiNZvc8Dp1jvS1ygV328rZl2FotvqEgz1X5qPfU96/Luldysa78y5C5n5I69beUlvS8WUKOC97sw1jKHmSRR50qMm2p7a/bR45fQsLNoCsNibJCEh058jwBrPlHWzo7Ltbl+0uESw2GEkvmo5V3f4LUIDpQJcERR05Scsy4jzvJ1Ypk7VpVtDAK8mNdI3RcVoHcwqOa5Vb15JeQXU0NgaWzT27aRHrtFcyyTgdT3pe7Mo0p1/WagijYnjy6hxXuEohH46Cew6V1LmSr1JYUYpfRvseOGSA7DFa3p+tTab0lJz1yDF1G5txh5wHUJ+0SRC6dvDUa9NJGp56kMCO8MgRH8m+RRcDexP8EWTXubvofYb68IZRIkYWxLhTV4wmqp5YvVqv6YYH3yq2D3hcc2XYRFPDMOUhOXZnkn7AgSTHIUZwNSpLo681dzkxq0NjheLggGTb3GGQIctx+LffdtbL6LNkjfHSRZumRyDRHzTspo3O/UpgEOrFA36LoUknASGtHX9du31xR/ZbxY+HcPC1l15MSADv30rtzq73rOVOPMuImItHE2P8rS63/3O7qNhAXN3/u127K17TPkVDOq7ezPHK4JyK96D+mDDl1RAkYuul0wh4Y/DBRmp/npLTK+TFgDzWfqIlniE0R2mlkOT2JRhYo50wOokly79asyGIc7sL9sBta3qSAkZ/aTXk32DA+5YVNlaq+8oELowkzBFpSztG7rMEcI+zeqv2wHpi6/EI6+uxWpPnGBQE1fxQ3nnbvYS8LKdy2AVXIoIMHpURBpomAmq0J/xaJsn+hJePNJ1WRMQSdB/g8kvaUeyDayCgbTIeu1YMGzKIXYd9P/tg2uBYr9iTQRvVl1wLhoxm52kL+UUUDAU3FHUGakvx81uoafaLUfBzRNUl35Og184LjTp0Bam632BP Kp/sVG7C 7pxhHsR2iNyqG4RHbsdJ8VKBnh8ZxRlGP8PPHiEmOTzoYsPoGeBj1JrbQKacihJ/1BUxKggh6/jolF3h2cPLPWRj3xGGVNxI3p72YMMKgZkY81zzwA5kU7R7wlzOHKjIcAXndrVhXYxNoFQZec4g/iuO+ikrKkXmQWvSGPYUNmxqU4zIu1b7x7MYfDr8NpgwJL+cc7OMZrPwI21yvcivEhKntycV5S+qoxlauNM5RGwKm+DsBh/Oy2SkNaYgvRwLdg3wIYUlb2MsRBWyU3BAmaolFniIqw/Qjbi2i+MT1LnClbQiQ7i6EW7L3bZbwTFIiBl3V33Bq6shc14tGxsmd3zdde7gczWDtuK2q+SG+NPGliYQ/yWh0P09M4wS1BgHHeHBbARLC9UuLhMJRn8fQ0g9jhm4aFS69bMA3rDbK7YKnUjWWj3xtAHQx3upMcwWKpnthq0todDIvNfKwoyxOqRWQBr0iAGp/UXCYyjEXNpve+90wKPMh/5qOWmJaBTxLTQrUI3pR/r5jrEWRd8tr3uALTpIodKy//xF9r/JQj52x/lTBW1knS7Y+zPeSzsy8OEVEZM0E+YwphuQgbpxgqnnqSA== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: During our experiment with zswap, we sometimes observe swap IOs due to occasional zswap store failures and writebacks-to-swap. These swapping IOs prevent many users who cannot tolerate swapping from adopting zswap to save memory and improve performance where possible. This patch adds the option to disable this behavior entirely: do not writeback to backing swapping device when a zswap store attempt fail, and do not write pages in the zswap pool back to the backing swap device (both when the pool is full, and when the new zswap shrinker is called). This new behavior can be opted-in/out on a per-cgroup basis via a new cgroup file. By default, writebacks to swap device is enabled, which is the previous behavior. Initially, writeback is enabled for the root cgroup, and a newly created cgroup will inherit the current setting of its parent. Note that this is subtly different from setting memory.swap.max to 0, as it still allows for pages to be stored in the zswap pool (which itself consumes swap space in its current form). This patch should be applied on top of the zswap shrinker series: https://lore.kernel.org/linux-mm/20231130194023.4102148-1-nphamcs@gmail.com/ as it also disables the zswap shrinker, a major source of zswap writebacks. Suggested-by: Johannes Weiner Signed-off-by: Nhat Pham Reviewed-by: Yosry Ahmed --- Documentation/admin-guide/cgroup-v2.rst | 12 ++++++++ Documentation/admin-guide/mm/zswap.rst | 6 ++++ include/linux/memcontrol.h | 12 ++++++++ include/linux/zswap.h | 6 ++++ mm/memcontrol.c | 38 +++++++++++++++++++++++++ mm/page_io.c | 6 ++++ mm/shmem.c | 3 +- mm/zswap.c | 13 +++++++-- 8 files changed, 92 insertions(+), 4 deletions(-) base-commit: cdcab2d34f129f593c0afbb2493bcaf41f4acd61 diff --git a/Documentation/admin-guide/cgroup-v2.rst b/Documentation/admin-guide/cgroup-v2.rst index 3f85254f3cef..2b4ac43efdc8 100644 --- a/Documentation/admin-guide/cgroup-v2.rst +++ b/Documentation/admin-guide/cgroup-v2.rst @@ -1679,6 +1679,18 @@ PAGE_SIZE multiple when read back. limit, it will refuse to take any more stores before existing entries fault back in or are written out to disk. + memory.zswap.writeback + A read-write single value file. The default value is "1". The + initial value of the root cgroup is 1, and when a new cgroup is + created, it inherits the current value of its parent. + + When this is set to 0, all swapping attempts to swapping devices + are disabled. This included both zswap writebacks, and swapping due + to zswap store failure. + + Note that this is subtly different from setting memory.swap.max to + 0, as it still allows for pages to be written to the zswap pool. + memory.pressure A read-only nested-keyed file. diff --git a/Documentation/admin-guide/mm/zswap.rst b/Documentation/admin-guide/mm/zswap.rst index 62fc244ec702..cfa653130346 100644 --- a/Documentation/admin-guide/mm/zswap.rst +++ b/Documentation/admin-guide/mm/zswap.rst @@ -153,6 +153,12 @@ attribute, e. g.:: Setting this parameter to 100 will disable the hysteresis. +Some users cannot tolerate the swapping that comes with zswap store failures +and zswap writebacks. Swapping can be disabled entirely (without disabling +zswap itself) on a cgroup-basis as follows: + + echo 0 > /sys/fs/cgroup//memory.zswap.writeback + When there is a sizable amount of cold memory residing in the zswap pool, it can be advantageous to proactively write these cold pages to swap and reclaim the memory for other use cases. By default, the zswap shrinker is disabled. diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h index 43b77363ab8e..5de775e6cdd9 100644 --- a/include/linux/memcontrol.h +++ b/include/linux/memcontrol.h @@ -219,6 +219,12 @@ struct mem_cgroup { #if defined(CONFIG_MEMCG_KMEM) && defined(CONFIG_ZSWAP) unsigned long zswap_max; + + /* + * Prevent pages from this memcg from being written back from zswap to + * swap, and from being swapped out on zswap store failures. + */ + bool zswap_writeback; #endif unsigned long soft_limit; @@ -1941,6 +1947,7 @@ static inline void count_objcg_event(struct obj_cgroup *objcg, bool obj_cgroup_may_zswap(struct obj_cgroup *objcg); void obj_cgroup_charge_zswap(struct obj_cgroup *objcg, size_t size); void obj_cgroup_uncharge_zswap(struct obj_cgroup *objcg, size_t size); +bool mem_cgroup_zswap_writeback_enabled(struct mem_cgroup *memcg); #else static inline bool obj_cgroup_may_zswap(struct obj_cgroup *objcg) { @@ -1954,6 +1961,11 @@ static inline void obj_cgroup_uncharge_zswap(struct obj_cgroup *objcg, size_t size) { } +static inline bool mem_cgroup_zswap_writeback_enabled(struct mem_cgroup *memcg) +{ + /* if zswap is disabled, do not block pages going to the swapping device */ + return true; +} #endif #endif /* _LINUX_MEMCONTROL_H */ diff --git a/include/linux/zswap.h b/include/linux/zswap.h index 08c240e16a01..a78ceaf3a65e 100644 --- a/include/linux/zswap.h +++ b/include/linux/zswap.h @@ -35,6 +35,7 @@ void zswap_swapoff(int type); void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg); void zswap_lruvec_state_init(struct lruvec *lruvec); void zswap_page_swapin(struct page *page); +bool is_zswap_enabled(void); #else struct zswap_lruvec_state {}; @@ -55,6 +56,11 @@ static inline void zswap_swapoff(int type) {} static inline void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg) {} static inline void zswap_lruvec_state_init(struct lruvec *lruvec) {} static inline void zswap_page_swapin(struct page *page) {} + +static inline bool is_zswap_enabled(void) +{ + return false; +} #endif #endif /* _LINUX_ZSWAP_H */ diff --git a/mm/memcontrol.c b/mm/memcontrol.c index d7bc47316acb..ae8c62c7aa53 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -5538,6 +5538,8 @@ mem_cgroup_css_alloc(struct cgroup_subsys_state *parent_css) WRITE_ONCE(memcg->soft_limit, PAGE_COUNTER_MAX); #if defined(CONFIG_MEMCG_KMEM) && defined(CONFIG_ZSWAP) memcg->zswap_max = PAGE_COUNTER_MAX; + WRITE_ONCE(memcg->zswap_writeback, + !parent || READ_ONCE(parent->zswap_writeback)); #endif page_counter_set_high(&memcg->swap, PAGE_COUNTER_MAX); if (parent) { @@ -8174,6 +8176,12 @@ void obj_cgroup_uncharge_zswap(struct obj_cgroup *objcg, size_t size) rcu_read_unlock(); } +bool mem_cgroup_zswap_writeback_enabled(struct mem_cgroup *memcg) +{ + /* if zswap is disabled, do not block pages going to the swapping device */ + return !is_zswap_enabled() || !memcg || READ_ONCE(memcg->zswap_writeback); +} + static u64 zswap_current_read(struct cgroup_subsys_state *css, struct cftype *cft) { @@ -8206,6 +8214,31 @@ static ssize_t zswap_max_write(struct kernfs_open_file *of, return nbytes; } +static int zswap_writeback_show(struct seq_file *m, void *v) +{ + struct mem_cgroup *memcg = mem_cgroup_from_seq(m); + + seq_printf(m, "%d\n", READ_ONCE(memcg->zswap_writeback)); + return 0; +} + +static ssize_t zswap_writeback_write(struct kernfs_open_file *of, + char *buf, size_t nbytes, loff_t off) +{ + struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); + int zswap_writeback; + ssize_t parse_ret = kstrtoint(strstrip(buf), 0, &zswap_writeback); + + if (parse_ret) + return parse_ret; + + if (zswap_writeback != 0 && zswap_writeback != 1) + return -EINVAL; + + WRITE_ONCE(memcg->zswap_writeback, zswap_writeback); + return nbytes; +} + static struct cftype zswap_files[] = { { .name = "zswap.current", @@ -8218,6 +8251,11 @@ static struct cftype zswap_files[] = { .seq_show = zswap_max_show, .write = zswap_max_write, }, + { + .name = "zswap.writeback", + .seq_show = zswap_writeback_show, + .write = zswap_writeback_write, + }, { } /* terminate */ }; #endif /* CONFIG_MEMCG_KMEM && CONFIG_ZSWAP */ diff --git a/mm/page_io.c b/mm/page_io.c index cb559ae324c6..5e606f1aa2f6 100644 --- a/mm/page_io.c +++ b/mm/page_io.c @@ -201,6 +201,12 @@ int swap_writepage(struct page *page, struct writeback_control *wbc) folio_end_writeback(folio); return 0; } + + if (!mem_cgroup_zswap_writeback_enabled(folio_memcg(folio))) { + folio_mark_dirty(folio); + return AOP_WRITEPAGE_ACTIVATE; + } + __swap_writepage(&folio->page, wbc); return 0; } diff --git a/mm/shmem.c b/mm/shmem.c index c62f904ba1ca..dd084fbafcf1 100644 --- a/mm/shmem.c +++ b/mm/shmem.c @@ -1514,8 +1514,7 @@ static int shmem_writepage(struct page *page, struct writeback_control *wbc) mutex_unlock(&shmem_swaplist_mutex); BUG_ON(folio_mapped(folio)); - swap_writepage(&folio->page, wbc); - return 0; + return swap_writepage(&folio->page, wbc); } mutex_unlock(&shmem_swaplist_mutex); diff --git a/mm/zswap.c b/mm/zswap.c index daaa949837f2..7ee54a3d8281 100644 --- a/mm/zswap.c +++ b/mm/zswap.c @@ -153,6 +153,11 @@ static bool zswap_shrinker_enabled = IS_ENABLED( CONFIG_ZSWAP_SHRINKER_DEFAULT_ON); module_param_named(shrinker_enabled, zswap_shrinker_enabled, bool, 0644); +bool is_zswap_enabled(void) +{ + return zswap_enabled; +} + /********************************* * data structures **********************************/ @@ -596,7 +601,8 @@ static unsigned long zswap_shrinker_scan(struct shrinker *shrinker, struct zswap_pool *pool = shrinker->private_data; bool encountered_page_in_swapcache = false; - if (!zswap_shrinker_enabled) { + if (!zswap_shrinker_enabled || + !mem_cgroup_zswap_writeback_enabled(sc->memcg)) { sc->nr_scanned = 0; return SHRINK_STOP; } @@ -637,7 +643,7 @@ static unsigned long zswap_shrinker_count(struct shrinker *shrinker, struct lruvec *lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(sc->nid)); unsigned long nr_backing, nr_stored, nr_freeable, nr_protected; - if (!zswap_shrinker_enabled) + if (!zswap_shrinker_enabled || !mem_cgroup_zswap_writeback_enabled(memcg)) return 0; #ifdef CONFIG_MEMCG_KMEM @@ -956,6 +962,9 @@ static int shrink_memcg(struct mem_cgroup *memcg) struct zswap_pool *pool; int nid, shrunk = 0; + if (!mem_cgroup_zswap_writeback_enabled(memcg)) + return -EINVAL; + /* * Skip zombies because their LRUs are reparented and we would be * reclaiming from the parent instead of the dead memcg.