Message ID | 20200608230819.832349-1-guro@fb.com (mailing list archive) |
---|---|
Headers | show |
Series | mm: memcg accounting of percpu memory | expand |
On Mon, Jun 08, 2020 at 04:08:14PM -0700, Roman Gushchin wrote: > This patchset adds percpu memory accounting to memory cgroups. > It's based on the rework of the slab controller and reuses concepts > and features introduced for the per-object slab accounting. > > Percpu memory is becoming more and more widely used by various > subsystems, and the total amount of memory controlled by the percpu > allocator can make a good part of the total memory. > > As an example, bpf maps can consume a lot of percpu memory, > and they are created by a user. Also, some cgroup internals > (e.g. memory controller statistics) can be quite large. > On a machine with many CPUs and big number of cgroups they > can consume hundreds of megabytes. > > So the lack of memcg accounting is creating a breach in the memory > isolation. Similar to the slab memory, percpu memory should be > accounted by default. > > Percpu allocations by their nature are scattered over multiple pages, > so they can't be tracked on the per-page basis. So the per-object > tracking introduced by the new slab controller is reused. > > The patchset implements charging of percpu allocations, adds > memcg-level statistics, enables accounting for percpu allocations made > by memory cgroup internals and provides some basic tests. > > To implement the accounting of percpu memory without a significant > memory and performance overhead the following approach is used: > all accounted allocations are placed into a separate percpu chunk > (or chunks). These chunks are similar to default chunks, except > that they do have an attached vector of pointers to obj_cgroup objects, > which is big enough to save a pointer for each allocated object. > On the allocation, if the allocation has to be accounted > (__GFP_ACCOUNT is passed, the allocating process belongs to a non-root > memory cgroup, etc), the memory cgroup is getting charged and if the maximum > limit is not exceeded the allocation is performed using a memcg-aware > chunk. Otherwise -ENOMEM is returned or the allocation is forced over > the limit, depending on gfp (as any other kernel memory allocation). > The memory cgroup information is saved in the obj_cgroup vector > at the corresponding offset. On the release time the memcg > information is restored from the vector and the cgroup is getting > uncharged. > Unaccounted allocations (at this point the absolute majority > of all percpu allocations) are performed in the old way, so no > additional overhead is expected. > > To avoid pinning dying memory cgroups by outstanding allocations, > obj_cgroup API is used instead of directly saving memory cgroup pointers. > obj_cgroup is basically a pointer to a memory cgroup with a standalone > reference counter. The trick is that it can be atomically swapped > to point at the parent cgroup, so that the original memory cgroup > can be released prior to all objects, which has been charged to it. > Because all charges and statistics are fully recursive, it's perfectly > correct to uncharge the parent cgroup instead. This scheme is used > in the slab memory accounting, and percpu memory can just follow > the scheme. > > This version is based on top of v6 of the new slab controller > patchset. The following patches are actually required by this series: > mm: memcg: factor out memcg- and lruvec-level changes out of __mod_lruvec_state() > mm: memcg: prepare for byte-sized vmstat items > mm: memcg: convert vmstat slab counters to bytes > mm: slub: implement SLUB version of obj_to_index() > mm: memcontrol: decouple reference counting from page accounting > mm: memcg/slab: obj_cgroup API Hello, Andrew! How this patchset should be routed: through the mm or percpu tree? It has been acked by Dennis (the percpu maintainer), but it does depend on first several patches from the slab controller rework patchset. The slab controller rework is ready to be merged: as in v6 most patches in the series were acked by Johannes and/or Vlastimil and no questions or concerns were raised after v6. Please, let me know if you want me to resend both patchsets. Thank you! Roman
On Tue, 16 Jun 2020 14:19:01 -0700 Roman Gushchin <guro@fb.com> wrote: > > > > This version is based on top of v6 of the new slab controller > > patchset. The following patches are actually required by this series: > > mm: memcg: factor out memcg- and lruvec-level changes out of __mod_lruvec_state() > > mm: memcg: prepare for byte-sized vmstat items > > mm: memcg: convert vmstat slab counters to bytes > > mm: slub: implement SLUB version of obj_to_index() > > mm: memcontrol: decouple reference counting from page accounting > > mm: memcg/slab: obj_cgroup API > > Hello, Andrew! > > How this patchset should be routed: through the mm or percpu tree? > > It has been acked by Dennis (the percpu maintainer), but it does depend > on first several patches from the slab controller rework patchset. I can grab both. > The slab controller rework is ready to be merged: as in v6 most patches > in the series were acked by Johannes and/or Vlastimil and no questions > or concerns were raised after v6. > > Please, let me know if you want me to resend both patchsets. There was quite a bit of valuable discussion in response to [0/n] which really should have been in the changelog[s] from day one. slab-vs-slub, performance testing, etc. So, umm, I'll take a look at both series now but I do think an enhanced [0/n] description is warranted?
On Wed, Jun 17, 2020 at 01:39:49PM -0700, Andrew Morton wrote: > On Tue, 16 Jun 2020 14:19:01 -0700 Roman Gushchin <guro@fb.com> wrote: > > > > > > > This version is based on top of v6 of the new slab controller > > > patchset. The following patches are actually required by this series: > > > mm: memcg: factor out memcg- and lruvec-level changes out of __mod_lruvec_state() > > > mm: memcg: prepare for byte-sized vmstat items > > > mm: memcg: convert vmstat slab counters to bytes > > > mm: slub: implement SLUB version of obj_to_index() > > > mm: memcontrol: decouple reference counting from page accounting > > > mm: memcg/slab: obj_cgroup API > > > > Hello, Andrew! > > > > How this patchset should be routed: through the mm or percpu tree? > > > > It has been acked by Dennis (the percpu maintainer), but it does depend > > on first several patches from the slab controller rework patchset. > > I can grab both. Perfect, thanks! > > > The slab controller rework is ready to be merged: as in v6 most patches > > in the series were acked by Johannes and/or Vlastimil and no questions > > or concerns were raised after v6. > > > > Please, let me know if you want me to resend both patchsets. > > There was quite a bit of valuable discussion in response to [0/n] which > really should have been in the changelog[s] from day one. > slab-vs-slub, performance testing, etc. > > So, umm, I'll take a look at both series now but I do think an enhanced > [0/n] description is warranted? > Yes, I'm running suggested tests right now, and will update on the results. Thanks!