diff mbox series

[v8,5/8] mm: zswap: Modify zswap_stored_pages to be atomic_long_t.

Message ID 20240928021620.8369-6-kanchana.p.sridhar@intel.com (mailing list archive)
State New
Headers show
Series mm: zswap swap-out of large folios | expand

Commit Message

Sridhar, Kanchana P Sept. 28, 2024, 2:16 a.m. UTC
For zswap_store() to support large folios, we need to be able to do
a batch update of zswap_stored_pages upon successful store of all pages
in the folio. For this, we need to add folio_nr_pages(), which returns
a long, to zswap_stored_pages.

Signed-off-by: Kanchana P Sridhar <kanchana.p.sridhar@intel.com>
---
 fs/proc/meminfo.c     |  2 +-
 include/linux/zswap.h |  2 +-
 mm/zswap.c            | 19 +++++++++++++------
 3 files changed, 15 insertions(+), 8 deletions(-)

Comments

Yosry Ahmed Sept. 28, 2024, 2:57 a.m. UTC | #1
On Fri, Sep 27, 2024 at 7:16 PM Kanchana P Sridhar
<kanchana.p.sridhar@intel.com> wrote:
>
> For zswap_store() to support large folios, we need to be able to do
> a batch update of zswap_stored_pages upon successful store of all pages
> in the folio. For this, we need to add folio_nr_pages(), which returns
> a long, to zswap_stored_pages.

Do we really need this? A lot of places in the kernel assign the
result of folio_nr_pages() to an int (thp_nr_pages(),
split_huge_pages_all(), etc). I don't think we need to worry about
folio_nr_pages() exceeding INT_MAX for a while.

>
> Signed-off-by: Kanchana P Sridhar <kanchana.p.sridhar@intel.com>
> ---
>  fs/proc/meminfo.c     |  2 +-
>  include/linux/zswap.h |  2 +-
>  mm/zswap.c            | 19 +++++++++++++------
>  3 files changed, 15 insertions(+), 8 deletions(-)
>
> diff --git a/fs/proc/meminfo.c b/fs/proc/meminfo.c
> index 245171d9164b..8ba9b1472390 100644
> --- a/fs/proc/meminfo.c
> +++ b/fs/proc/meminfo.c
> @@ -91,7 +91,7 @@ static int meminfo_proc_show(struct seq_file *m, void *v)
>  #ifdef CONFIG_ZSWAP
>         show_val_kb(m, "Zswap:          ", zswap_total_pages());
>         seq_printf(m,  "Zswapped:       %8lu kB\n",
> -                  (unsigned long)atomic_read(&zswap_stored_pages) <<
> +                  (unsigned long)atomic_long_read(&zswap_stored_pages) <<
>                    (PAGE_SHIFT - 10));
>  #endif
>         show_val_kb(m, "Dirty:          ",
> diff --git a/include/linux/zswap.h b/include/linux/zswap.h
> index 9cd1beef0654..d961ead91bf1 100644
> --- a/include/linux/zswap.h
> +++ b/include/linux/zswap.h
> @@ -7,7 +7,7 @@
>
>  struct lruvec;
>
> -extern atomic_t zswap_stored_pages;
> +extern atomic_long_t zswap_stored_pages;
>
>  #ifdef CONFIG_ZSWAP
>
> diff --git a/mm/zswap.c b/mm/zswap.c
> index 0f281e50a034..43e4e216db41 100644
> --- a/mm/zswap.c
> +++ b/mm/zswap.c
> @@ -43,7 +43,7 @@
>  * statistics
>  **********************************/
>  /* The number of compressed pages currently stored in zswap */
> -atomic_t zswap_stored_pages = ATOMIC_INIT(0);
> +atomic_long_t zswap_stored_pages = ATOMIC_INIT(0);
>
>  /*
>   * The statistics below are not protected from concurrent access for
> @@ -802,7 +802,7 @@ static void zswap_entry_free(struct zswap_entry *entry)
>                 obj_cgroup_put(entry->objcg);
>         }
>         zswap_entry_cache_free(entry);
> -       atomic_dec(&zswap_stored_pages);
> +       atomic_long_dec(&zswap_stored_pages);
>  }
>
>  /*********************************
> @@ -1232,7 +1232,7 @@ static unsigned long zswap_shrinker_count(struct shrinker *shrinker,
>                 nr_stored = memcg_page_state(memcg, MEMCG_ZSWAPPED);
>         } else {
>                 nr_backing = zswap_total_pages();
> -               nr_stored = atomic_read(&zswap_stored_pages);
> +               nr_stored = atomic_long_read(&zswap_stored_pages);
>         }
>
>         if (!nr_stored)
> @@ -1501,7 +1501,7 @@ bool zswap_store(struct folio *folio)
>         }
>
>         /* update stats */
> -       atomic_inc(&zswap_stored_pages);
> +       atomic_long_inc(&zswap_stored_pages);
>         count_vm_event(ZSWPOUT);
>
>         return true;
> @@ -1650,6 +1650,13 @@ static int debugfs_get_total_size(void *data, u64 *val)
>  }
>  DEFINE_DEBUGFS_ATTRIBUTE(total_size_fops, debugfs_get_total_size, NULL, "%llu\n");
>
> +static int debugfs_get_stored_pages(void *data, u64 *val)
> +{
> +       *val = atomic_long_read(&zswap_stored_pages);
> +       return 0;
> +}
> +DEFINE_DEBUGFS_ATTRIBUTE(stored_pages_fops, debugfs_get_stored_pages, NULL, "%llu\n");
> +
>  static int zswap_debugfs_init(void)
>  {
>         if (!debugfs_initialized())
> @@ -1673,8 +1680,8 @@ static int zswap_debugfs_init(void)
>                            zswap_debugfs_root, &zswap_written_back_pages);
>         debugfs_create_file("pool_total_size", 0444,
>                             zswap_debugfs_root, NULL, &total_size_fops);
> -       debugfs_create_atomic_t("stored_pages", 0444,
> -                               zswap_debugfs_root, &zswap_stored_pages);
> +       debugfs_create_file("stored_pages", 0444,
> +                           zswap_debugfs_root, NULL, &stored_pages_fops);
>
>         return 0;
>  }
> --
> 2.27.0
>
Matthew Wilcox (Oracle) Sept. 28, 2024, 4:50 a.m. UTC | #2
On Fri, Sep 27, 2024 at 07:57:49PM -0700, Yosry Ahmed wrote:
> On Fri, Sep 27, 2024 at 7:16 PM Kanchana P Sridhar
> <kanchana.p.sridhar@intel.com> wrote:
> >
> > For zswap_store() to support large folios, we need to be able to do
> > a batch update of zswap_stored_pages upon successful store of all pages
> > in the folio. For this, we need to add folio_nr_pages(), which returns
> > a long, to zswap_stored_pages.
> 
> Do we really need this? A lot of places in the kernel assign the
> result of folio_nr_pages() to an int (thp_nr_pages(),
> split_huge_pages_all(), etc). I don't think we need to worry about
> folio_nr_pages() exceeding INT_MAX for a while.

You'd be surprised.  Let's assume we add support for PUD-sized pages
(personally I think this is too large to make sense, but some people can't
be told).  On arm64, we can have a 64kB page size, so that's 13 bits per
level for a total of 2^26 pages per PUD.  That feels uncomfortable close
to 2^32 to me.

Anywhere you've found that's using an int to store folio_nr_pages() is
somewhere we should probably switch to long.  And this, btw, is why I've
moved from using an int to store folio_size() to using size_t.  A PMD is
already 512MB (with a 64KB page size), and so a PUD will be 4TB.

thp_nr_pages() is not a good example.  I'll be happy when we kill it;
we're actually almost there.
Yosry Ahmed Sept. 28, 2024, 8:12 a.m. UTC | #3
On Fri, Sep 27, 2024 at 9:51 PM Matthew Wilcox <willy@infradead.org> wrote:
>
> On Fri, Sep 27, 2024 at 07:57:49PM -0700, Yosry Ahmed wrote:
> > On Fri, Sep 27, 2024 at 7:16 PM Kanchana P Sridhar
> > <kanchana.p.sridhar@intel.com> wrote:
> > >
> > > For zswap_store() to support large folios, we need to be able to do
> > > a batch update of zswap_stored_pages upon successful store of all pages
> > > in the folio. For this, we need to add folio_nr_pages(), which returns
> > > a long, to zswap_stored_pages.
> >
> > Do we really need this? A lot of places in the kernel assign the
> > result of folio_nr_pages() to an int (thp_nr_pages(),
> > split_huge_pages_all(), etc). I don't think we need to worry about
> > folio_nr_pages() exceeding INT_MAX for a while.
>
> You'd be surprised.  Let's assume we add support for PUD-sized pages
> (personally I think this is too large to make sense, but some people can't
> be told).  On arm64, we can have a 64kB page size, so that's 13 bits per
> level for a total of 2^26 pages per PUD.  That feels uncomfortable close
> to 2^32 to me.
>
> Anywhere you've found that's using an int to store folio_nr_pages() is
> somewhere we should probably switch to long.

There are a lot of them: rmap.c, shmem.c, khugepaged.c, etc.

> And this, btw, is why I've
> moved from using an int to store folio_size() to using size_t.  A PMD is
> already 512MB (with a 64KB page size), and so a PUD will be 4TB.

Thanks for pointing this out. I assumed the presence of many places
using int to store folio_nr_pages() means that it's a general
assumption.

Also, if we think it's possible that a single folio size may approach
INT_MAX, then we are in bigger trouble for zswap_stored_pages, because
that's the total number of pages stored in zswap on the entire system.
That's much more likely to exceed INT_MAX than a single folio.

>
> thp_nr_pages() is not a good example.  I'll be happy when we kill it;
> we're actually almost there.

Yeah I can only see 2 callers.
Yosry Ahmed Sept. 28, 2024, 8:13 a.m. UTC | #4
On Fri, Sep 27, 2024 at 7:16 PM Kanchana P Sridhar
<kanchana.p.sridhar@intel.com> wrote:
>
> For zswap_store() to support large folios, we need to be able to do
> a batch update of zswap_stored_pages upon successful store of all pages
> in the folio. For this, we need to add folio_nr_pages(), which returns
> a long, to zswap_stored_pages.
>
> Signed-off-by: Kanchana P Sridhar <kanchana.p.sridhar@intel.com>
> ---
>  fs/proc/meminfo.c     |  2 +-
>  include/linux/zswap.h |  2 +-
>  mm/zswap.c            | 19 +++++++++++++------
>  3 files changed, 15 insertions(+), 8 deletions(-)
>
> diff --git a/fs/proc/meminfo.c b/fs/proc/meminfo.c
> index 245171d9164b..8ba9b1472390 100644
> --- a/fs/proc/meminfo.c
> +++ b/fs/proc/meminfo.c
> @@ -91,7 +91,7 @@ static int meminfo_proc_show(struct seq_file *m, void *v)
>  #ifdef CONFIG_ZSWAP
>         show_val_kb(m, "Zswap:          ", zswap_total_pages());
>         seq_printf(m,  "Zswapped:       %8lu kB\n",
> -                  (unsigned long)atomic_read(&zswap_stored_pages) <<
> +                  (unsigned long)atomic_long_read(&zswap_stored_pages) <<

Do we still need this cast? "HardwareCorrupted" seems to be using
atomic_long_read() without a cast.

Otherwise this LGTM:
Acked-by: Yosry Ahmed <yosryahmed@google.com>

>                    (PAGE_SHIFT - 10));
>  #endif
>         show_val_kb(m, "Dirty:          ",
> diff --git a/include/linux/zswap.h b/include/linux/zswap.h
> index 9cd1beef0654..d961ead91bf1 100644
> --- a/include/linux/zswap.h
> +++ b/include/linux/zswap.h
> @@ -7,7 +7,7 @@
>
>  struct lruvec;
>
> -extern atomic_t zswap_stored_pages;
> +extern atomic_long_t zswap_stored_pages;
>
>  #ifdef CONFIG_ZSWAP
>
> diff --git a/mm/zswap.c b/mm/zswap.c
> index 0f281e50a034..43e4e216db41 100644
> --- a/mm/zswap.c
> +++ b/mm/zswap.c
> @@ -43,7 +43,7 @@
>  * statistics
>  **********************************/
>  /* The number of compressed pages currently stored in zswap */
> -atomic_t zswap_stored_pages = ATOMIC_INIT(0);
> +atomic_long_t zswap_stored_pages = ATOMIC_INIT(0);
>
>  /*
>   * The statistics below are not protected from concurrent access for
> @@ -802,7 +802,7 @@ static void zswap_entry_free(struct zswap_entry *entry)
>                 obj_cgroup_put(entry->objcg);
>         }
>         zswap_entry_cache_free(entry);
> -       atomic_dec(&zswap_stored_pages);
> +       atomic_long_dec(&zswap_stored_pages);
>  }
>
>  /*********************************
> @@ -1232,7 +1232,7 @@ static unsigned long zswap_shrinker_count(struct shrinker *shrinker,
>                 nr_stored = memcg_page_state(memcg, MEMCG_ZSWAPPED);
>         } else {
>                 nr_backing = zswap_total_pages();
> -               nr_stored = atomic_read(&zswap_stored_pages);
> +               nr_stored = atomic_long_read(&zswap_stored_pages);
>         }
>
>         if (!nr_stored)
> @@ -1501,7 +1501,7 @@ bool zswap_store(struct folio *folio)
>         }
>
>         /* update stats */
> -       atomic_inc(&zswap_stored_pages);
> +       atomic_long_inc(&zswap_stored_pages);
>         count_vm_event(ZSWPOUT);
>
>         return true;
> @@ -1650,6 +1650,13 @@ static int debugfs_get_total_size(void *data, u64 *val)
>  }
>  DEFINE_DEBUGFS_ATTRIBUTE(total_size_fops, debugfs_get_total_size, NULL, "%llu\n");
>
> +static int debugfs_get_stored_pages(void *data, u64 *val)
> +{
> +       *val = atomic_long_read(&zswap_stored_pages);
> +       return 0;
> +}
> +DEFINE_DEBUGFS_ATTRIBUTE(stored_pages_fops, debugfs_get_stored_pages, NULL, "%llu\n");
> +
>  static int zswap_debugfs_init(void)
>  {
>         if (!debugfs_initialized())
> @@ -1673,8 +1680,8 @@ static int zswap_debugfs_init(void)
>                            zswap_debugfs_root, &zswap_written_back_pages);
>         debugfs_create_file("pool_total_size", 0444,
>                             zswap_debugfs_root, NULL, &total_size_fops);
> -       debugfs_create_atomic_t("stored_pages", 0444,
> -                               zswap_debugfs_root, &zswap_stored_pages);
> +       debugfs_create_file("stored_pages", 0444,
> +                           zswap_debugfs_root, NULL, &stored_pages_fops);
>
>         return 0;
>  }
> --
> 2.27.0
>
Johannes Weiner Sept. 28, 2024, 1:53 p.m. UTC | #5
On Fri, Sep 27, 2024 at 07:16:17PM -0700, Kanchana P Sridhar wrote:
> For zswap_store() to support large folios, we need to be able to do
> a batch update of zswap_stored_pages upon successful store of all pages
> in the folio. For this, we need to add folio_nr_pages(), which returns
> a long, to zswap_stored_pages.
> 
> Signed-off-by: Kanchana P Sridhar <kanchana.p.sridhar@intel.com>

Long for pages makes sense to me even independent of the large folios
coming in. An int is just 8TB in 4k (base) pages.

Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Nhat Pham Sept. 28, 2024, 11:27 p.m. UTC | #6
On Fri, Sep 27, 2024 at 7:16 PM Kanchana P Sridhar
<kanchana.p.sridhar@intel.com> wrote:
>
> For zswap_store() to support large folios, we need to be able to do
> a batch update of zswap_stored_pages upon successful store of all pages
> in the folio. For this, we need to add folio_nr_pages(), which returns
> a long, to zswap_stored_pages.
>
> Signed-off-by: Kanchana P Sridhar <kanchana.p.sridhar@intel.com>

Reviewed-by: Nhat Pham <nphamcs@gmail.com>
Sridhar, Kanchana P Sept. 29, 2024, 9:03 p.m. UTC | #7
> -----Original Message-----
> From: Johannes Weiner <hannes@cmpxchg.org>
> Sent: Saturday, September 28, 2024 6:54 AM
> To: Sridhar, Kanchana P <kanchana.p.sridhar@intel.com>
> Cc: linux-kernel@vger.kernel.org; linux-mm@kvack.org;
> yosryahmed@google.com; nphamcs@gmail.com;
> chengming.zhou@linux.dev; usamaarif642@gmail.com;
> shakeel.butt@linux.dev; ryan.roberts@arm.com; Huang, Ying
> <ying.huang@intel.com>; 21cnbao@gmail.com; akpm@linux-foundation.org;
> Zou, Nanhai <nanhai.zou@intel.com>; Feghali, Wajdi K
> <wajdi.k.feghali@intel.com>; Gopal, Vinodh <vinodh.gopal@intel.com>
> Subject: Re: [PATCH v8 5/8] mm: zswap: Modify zswap_stored_pages to be
> atomic_long_t.
> 
> On Fri, Sep 27, 2024 at 07:16:17PM -0700, Kanchana P Sridhar wrote:
> > For zswap_store() to support large folios, we need to be able to do
> > a batch update of zswap_stored_pages upon successful store of all pages
> > in the folio. For this, we need to add folio_nr_pages(), which returns
> > a long, to zswap_stored_pages.
> >
> > Signed-off-by: Kanchana P Sridhar <kanchana.p.sridhar@intel.com>
> 
> Long for pages makes sense to me even independent of the large folios
> coming in. An int is just 8TB in 4k (base) pages.
> 
> Acked-by: Johannes Weiner <hannes@cmpxchg.org>

Thanks Johannes for the Acked-by's!

Thanks,
Kanchana
Sridhar, Kanchana P Sept. 29, 2024, 9:04 p.m. UTC | #8
> -----Original Message-----
> From: Yosry Ahmed <yosryahmed@google.com>
> Sent: Saturday, September 28, 2024 1:13 AM
> To: Sridhar, Kanchana P <kanchana.p.sridhar@intel.com>
> Cc: linux-kernel@vger.kernel.org; linux-mm@kvack.org;
> hannes@cmpxchg.org; nphamcs@gmail.com; chengming.zhou@linux.dev;
> usamaarif642@gmail.com; shakeel.butt@linux.dev; ryan.roberts@arm.com;
> Huang, Ying <ying.huang@intel.com>; 21cnbao@gmail.com; akpm@linux-
> foundation.org; Zou, Nanhai <nanhai.zou@intel.com>; Feghali, Wajdi K
> <wajdi.k.feghali@intel.com>; Gopal, Vinodh <vinodh.gopal@intel.com>
> Subject: Re: [PATCH v8 5/8] mm: zswap: Modify zswap_stored_pages to be
> atomic_long_t.
> 
> On Fri, Sep 27, 2024 at 7:16 PM Kanchana P Sridhar
> <kanchana.p.sridhar@intel.com> wrote:
> >
> > For zswap_store() to support large folios, we need to be able to do
> > a batch update of zswap_stored_pages upon successful store of all pages
> > in the folio. For this, we need to add folio_nr_pages(), which returns
> > a long, to zswap_stored_pages.
> >
> > Signed-off-by: Kanchana P Sridhar <kanchana.p.sridhar@intel.com>
> > ---
> >  fs/proc/meminfo.c     |  2 +-
> >  include/linux/zswap.h |  2 +-
> >  mm/zswap.c            | 19 +++++++++++++------
> >  3 files changed, 15 insertions(+), 8 deletions(-)
> >
> > diff --git a/fs/proc/meminfo.c b/fs/proc/meminfo.c
> > index 245171d9164b..8ba9b1472390 100644
> > --- a/fs/proc/meminfo.c
> > +++ b/fs/proc/meminfo.c
> > @@ -91,7 +91,7 @@ static int meminfo_proc_show(struct seq_file *m,
> void *v)
> >  #ifdef CONFIG_ZSWAP
> >         show_val_kb(m, "Zswap:          ", zswap_total_pages());
> >         seq_printf(m,  "Zswapped:       %8lu kB\n",
> > -                  (unsigned long)atomic_read(&zswap_stored_pages) <<
> > +                  (unsigned long)atomic_long_read(&zswap_stored_pages) <<
> 
> Do we still need this cast? "HardwareCorrupted" seems to be using
> atomic_long_read() without a cast.
> 
> Otherwise this LGTM:
> Acked-by: Yosry Ahmed <yosryahmed@google.com>

Thanks Yosry for the Acked-by's!

Thanks,
Kanchana

> 
> >                    (PAGE_SHIFT - 10));
> >  #endif
> >         show_val_kb(m, "Dirty:          ",
> > diff --git a/include/linux/zswap.h b/include/linux/zswap.h
> > index 9cd1beef0654..d961ead91bf1 100644
> > --- a/include/linux/zswap.h
> > +++ b/include/linux/zswap.h
> > @@ -7,7 +7,7 @@
> >
> >  struct lruvec;
> >
> > -extern atomic_t zswap_stored_pages;
> > +extern atomic_long_t zswap_stored_pages;
> >
> >  #ifdef CONFIG_ZSWAP
> >
> > diff --git a/mm/zswap.c b/mm/zswap.c
> > index 0f281e50a034..43e4e216db41 100644
> > --- a/mm/zswap.c
> > +++ b/mm/zswap.c
> > @@ -43,7 +43,7 @@
> >  * statistics
> >  **********************************/
> >  /* The number of compressed pages currently stored in zswap */
> > -atomic_t zswap_stored_pages = ATOMIC_INIT(0);
> > +atomic_long_t zswap_stored_pages = ATOMIC_INIT(0);
> >
> >  /*
> >   * The statistics below are not protected from concurrent access for
> > @@ -802,7 +802,7 @@ static void zswap_entry_free(struct zswap_entry
> *entry)
> >                 obj_cgroup_put(entry->objcg);
> >         }
> >         zswap_entry_cache_free(entry);
> > -       atomic_dec(&zswap_stored_pages);
> > +       atomic_long_dec(&zswap_stored_pages);
> >  }
> >
> >  /*********************************
> > @@ -1232,7 +1232,7 @@ static unsigned long zswap_shrinker_count(struct
> shrinker *shrinker,
> >                 nr_stored = memcg_page_state(memcg, MEMCG_ZSWAPPED);
> >         } else {
> >                 nr_backing = zswap_total_pages();
> > -               nr_stored = atomic_read(&zswap_stored_pages);
> > +               nr_stored = atomic_long_read(&zswap_stored_pages);
> >         }
> >
> >         if (!nr_stored)
> > @@ -1501,7 +1501,7 @@ bool zswap_store(struct folio *folio)
> >         }
> >
> >         /* update stats */
> > -       atomic_inc(&zswap_stored_pages);
> > +       atomic_long_inc(&zswap_stored_pages);
> >         count_vm_event(ZSWPOUT);
> >
> >         return true;
> > @@ -1650,6 +1650,13 @@ static int debugfs_get_total_size(void *data,
> u64 *val)
> >  }
> >  DEFINE_DEBUGFS_ATTRIBUTE(total_size_fops, debugfs_get_total_size,
> NULL, "%llu\n");
> >
> > +static int debugfs_get_stored_pages(void *data, u64 *val)
> > +{
> > +       *val = atomic_long_read(&zswap_stored_pages);
> > +       return 0;
> > +}
> > +DEFINE_DEBUGFS_ATTRIBUTE(stored_pages_fops,
> debugfs_get_stored_pages, NULL, "%llu\n");
> > +
> >  static int zswap_debugfs_init(void)
> >  {
> >         if (!debugfs_initialized())
> > @@ -1673,8 +1680,8 @@ static int zswap_debugfs_init(void)
> >                            zswap_debugfs_root, &zswap_written_back_pages);
> >         debugfs_create_file("pool_total_size", 0444,
> >                             zswap_debugfs_root, NULL, &total_size_fops);
> > -       debugfs_create_atomic_t("stored_pages", 0444,
> > -                               zswap_debugfs_root, &zswap_stored_pages);
> > +       debugfs_create_file("stored_pages", 0444,
> > +                           zswap_debugfs_root, NULL, &stored_pages_fops);
> >
> >         return 0;
> >  }
> > --
> > 2.27.0
> >
diff mbox series

Patch

diff --git a/fs/proc/meminfo.c b/fs/proc/meminfo.c
index 245171d9164b..8ba9b1472390 100644
--- a/fs/proc/meminfo.c
+++ b/fs/proc/meminfo.c
@@ -91,7 +91,7 @@  static int meminfo_proc_show(struct seq_file *m, void *v)
 #ifdef CONFIG_ZSWAP
 	show_val_kb(m, "Zswap:          ", zswap_total_pages());
 	seq_printf(m,  "Zswapped:       %8lu kB\n",
-		   (unsigned long)atomic_read(&zswap_stored_pages) <<
+		   (unsigned long)atomic_long_read(&zswap_stored_pages) <<
 		   (PAGE_SHIFT - 10));
 #endif
 	show_val_kb(m, "Dirty:          ",
diff --git a/include/linux/zswap.h b/include/linux/zswap.h
index 9cd1beef0654..d961ead91bf1 100644
--- a/include/linux/zswap.h
+++ b/include/linux/zswap.h
@@ -7,7 +7,7 @@ 
 
 struct lruvec;
 
-extern atomic_t zswap_stored_pages;
+extern atomic_long_t zswap_stored_pages;
 
 #ifdef CONFIG_ZSWAP
 
diff --git a/mm/zswap.c b/mm/zswap.c
index 0f281e50a034..43e4e216db41 100644
--- a/mm/zswap.c
+++ b/mm/zswap.c
@@ -43,7 +43,7 @@ 
 * statistics
 **********************************/
 /* The number of compressed pages currently stored in zswap */
-atomic_t zswap_stored_pages = ATOMIC_INIT(0);
+atomic_long_t zswap_stored_pages = ATOMIC_INIT(0);
 
 /*
  * The statistics below are not protected from concurrent access for
@@ -802,7 +802,7 @@  static void zswap_entry_free(struct zswap_entry *entry)
 		obj_cgroup_put(entry->objcg);
 	}
 	zswap_entry_cache_free(entry);
-	atomic_dec(&zswap_stored_pages);
+	atomic_long_dec(&zswap_stored_pages);
 }
 
 /*********************************
@@ -1232,7 +1232,7 @@  static unsigned long zswap_shrinker_count(struct shrinker *shrinker,
 		nr_stored = memcg_page_state(memcg, MEMCG_ZSWAPPED);
 	} else {
 		nr_backing = zswap_total_pages();
-		nr_stored = atomic_read(&zswap_stored_pages);
+		nr_stored = atomic_long_read(&zswap_stored_pages);
 	}
 
 	if (!nr_stored)
@@ -1501,7 +1501,7 @@  bool zswap_store(struct folio *folio)
 	}
 
 	/* update stats */
-	atomic_inc(&zswap_stored_pages);
+	atomic_long_inc(&zswap_stored_pages);
 	count_vm_event(ZSWPOUT);
 
 	return true;
@@ -1650,6 +1650,13 @@  static int debugfs_get_total_size(void *data, u64 *val)
 }
 DEFINE_DEBUGFS_ATTRIBUTE(total_size_fops, debugfs_get_total_size, NULL, "%llu\n");
 
+static int debugfs_get_stored_pages(void *data, u64 *val)
+{
+	*val = atomic_long_read(&zswap_stored_pages);
+	return 0;
+}
+DEFINE_DEBUGFS_ATTRIBUTE(stored_pages_fops, debugfs_get_stored_pages, NULL, "%llu\n");
+
 static int zswap_debugfs_init(void)
 {
 	if (!debugfs_initialized())
@@ -1673,8 +1680,8 @@  static int zswap_debugfs_init(void)
 			   zswap_debugfs_root, &zswap_written_back_pages);
 	debugfs_create_file("pool_total_size", 0444,
 			    zswap_debugfs_root, NULL, &total_size_fops);
-	debugfs_create_atomic_t("stored_pages", 0444,
-				zswap_debugfs_root, &zswap_stored_pages);
+	debugfs_create_file("stored_pages", 0444,
+			    zswap_debugfs_root, NULL, &stored_pages_fops);
 
 	return 0;
 }