Message ID | 20240809073309.2134488-4-kees@kernel.org (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | slab: Allocate and use per-call-site caches | expand |
On Fri, Aug 9, 2024 at 12:33 AM Kees Cook <kees@kernel.org> wrote: > > For slab allocations, record whether the call site is using a fixed > size (i.e. compile time constant) or a dynamic size. Report the results > in /proc/allocinfo. > > Improvements needed: > - examine realloc routines for needed coverage > > Signed-off-by: Kees Cook <kees@kernel.org> > --- > Cc: Suren Baghdasaryan <surenb@google.com> > Cc: Kent Overstreet <kent.overstreet@linux.dev> > Cc: Vlastimil Babka <vbabka@suse.cz> > Cc: Christoph Lameter <cl@linux.com> > Cc: Pekka Enberg <penberg@kernel.org> > Cc: David Rientjes <rientjes@google.com> > Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com> > Cc: Andrew Morton <akpm@linux-foundation.org> > Cc: Roman Gushchin <roman.gushchin@linux.dev> > Cc: Hyeonggon Yoo <42.hyeyoo@gmail.com> > Cc: linux-mm@kvack.org > --- > include/linux/alloc_tag.h | 30 ++++++++++++++++++++++++++---- > include/linux/slab.h | 16 ++++++++-------- > lib/alloc_tag.c | 8 ++++++++ > mm/Kconfig | 8 ++++++++ > 4 files changed, 50 insertions(+), 12 deletions(-) > > diff --git a/include/linux/alloc_tag.h b/include/linux/alloc_tag.h > index 8c61ccd161ba..f5d8c5849b82 100644 > --- a/include/linux/alloc_tag.h > +++ b/include/linux/alloc_tag.h > @@ -20,6 +20,19 @@ struct alloc_tag_counters { > u64 calls; > }; > > +#ifdef CONFIG_SLAB_PER_SITE > +struct alloc_meta { > + /* 0 means non-slab, SIZE_MAX means dynamic, and everything else is fixed-size. */ > + size_t sized; > +}; > +#define ALLOC_META_INIT(_size) { \ > + .sized = (__builtin_constant_p(_size) ? (_size) : SIZE_MAX), \ > + } > +#else > +struct alloc_meta { }; > +#define ALLOC_META_INIT(_size) { } > +#endif > + > /* > * An instance of this structure is created in a special ELF section at every > * allocation callsite. At runtime, the special section is treated as > @@ -27,6 +40,7 @@ struct alloc_tag_counters { > */ > struct alloc_tag { > struct codetag ct; > + struct alloc_meta meta; > struct alloc_tag_counters __percpu *counters; > } __aligned(8); > > @@ -74,19 +88,21 @@ static inline struct alloc_tag *ct_to_alloc_tag(struct codetag *ct) > */ > DECLARE_PER_CPU(struct alloc_tag_counters, _shared_alloc_tag); > > -#define DEFINE_ALLOC_TAG(_alloc_tag) \ > +#define DEFINE_ALLOC_TAG(_alloc_tag, _meta_init) \ > static struct alloc_tag _alloc_tag __used __aligned(8) \ > __section("alloc_tags") = { \ > .ct = CODE_TAG_INIT, \ > + .meta = _meta_init, \ > .counters = &_shared_alloc_tag }; > > #else /* ARCH_NEEDS_WEAK_PER_CPU */ > > -#define DEFINE_ALLOC_TAG(_alloc_tag) \ > +#define DEFINE_ALLOC_TAG(_alloc_tag, _meta_init) \ > static DEFINE_PER_CPU(struct alloc_tag_counters, _alloc_tag_cntr); \ > static struct alloc_tag _alloc_tag __used __aligned(8) \ > __section("alloc_tags") = { \ > .ct = CODE_TAG_INIT, \ > + .meta = _meta_init, \ > .counters = &_alloc_tag_cntr }; > > #endif /* ARCH_NEEDS_WEAK_PER_CPU */ > @@ -191,7 +207,7 @@ static inline void alloc_tag_sub(union codetag_ref *ref, size_t bytes) > > #else /* CONFIG_MEM_ALLOC_PROFILING */ > > -#define DEFINE_ALLOC_TAG(_alloc_tag) > +#define DEFINE_ALLOC_TAG(_alloc_tag, _meta_init) > static inline bool mem_alloc_profiling_enabled(void) { return false; } > static inline void alloc_tag_add(union codetag_ref *ref, struct alloc_tag *tag, > size_t bytes) {} > @@ -210,8 +226,14 @@ static inline void alloc_tag_sub(union codetag_ref *ref, size_t bytes) {} > > #define alloc_hooks(_do_alloc) \ > ({ \ > - DEFINE_ALLOC_TAG(_alloc_tag); \ > + DEFINE_ALLOC_TAG(_alloc_tag, { }); \ > alloc_hooks_tag(&_alloc_tag, _do_alloc); \ > }) > > +#define alloc_sized_hooks(_do_alloc, _size, ...) \ > +({ \ > + DEFINE_ALLOC_TAG(_alloc_tag, ALLOC_META_INIT(_size)); \ > + alloc_hooks_tag(&_alloc_tag, _do_alloc(_size, __VA_ARGS__)); \ > +}) > + > #endif /* _LINUX_ALLOC_TAG_H */ > diff --git a/include/linux/slab.h b/include/linux/slab.h > index 86cb61a0102c..314d24c79e05 100644 > --- a/include/linux/slab.h > +++ b/include/linux/slab.h > @@ -541,7 +541,7 @@ static_assert(PAGE_SHIFT <= 20); > */ > void *kmem_cache_alloc_noprof(struct kmem_cache *cachep, > gfp_t flags) __assume_slab_alignment __malloc; > -#define kmem_cache_alloc(...) alloc_hooks(kmem_cache_alloc_noprof(__VA_ARGS__)) > +#define kmem_cache_alloc(...) alloc_hooks(kmem_cache_alloc_noprof(__VA_ARGS__)) nit: seems like an unnecessary churn. > > void *kmem_cache_alloc_lru_noprof(struct kmem_cache *s, struct list_lru *lru, > gfp_t gfpflags) __assume_slab_alignment __malloc; > @@ -685,7 +685,7 @@ static __always_inline __alloc_size(1) void *kmalloc_noprof(size_t size, gfp_t f > } > return __kmalloc_noprof(size, flags); > } > -#define kmalloc(...) alloc_hooks(kmalloc_noprof(__VA_ARGS__)) > +#define kmalloc(size, ...) alloc_sized_hooks(kmalloc_noprof, size, __VA_ARGS__) > > #define kmem_buckets_alloc(_b, _size, _flags) \ > alloc_hooks(__kmalloc_node_noprof(PASS_BUCKET_PARAMS(_size, _b), _flags, NUMA_NO_NODE)) > @@ -708,7 +708,7 @@ static __always_inline __alloc_size(1) void *kmalloc_node_noprof(size_t size, gf > } > return __kmalloc_node_noprof(PASS_BUCKET_PARAMS(size, NULL), flags, node); > } > -#define kmalloc_node(...) alloc_hooks(kmalloc_node_noprof(__VA_ARGS__)) > +#define kmalloc_node(size, ...) alloc_sized_hooks(kmalloc_node_noprof, size, __VA_ARGS__) > > /** > * kmalloc_array - allocate memory for an array. > @@ -726,7 +726,7 @@ static inline __alloc_size(1, 2) void *kmalloc_array_noprof(size_t n, size_t siz > return kmalloc_noprof(bytes, flags); > return kmalloc_noprof(bytes, flags); > } > -#define kmalloc_array(...) alloc_hooks(kmalloc_array_noprof(__VA_ARGS__)) > +#define kmalloc_array(...) alloc_hooks(kmalloc_array_noprof(__VA_ARGS__)) ditto. > > /** > * krealloc_array - reallocate memory for an array. > @@ -761,8 +761,8 @@ void *__kmalloc_node_track_caller_noprof(DECL_BUCKET_PARAMS(size, b), gfp_t flag > unsigned long caller) __alloc_size(1); > #define kmalloc_node_track_caller_noprof(size, flags, node, caller) \ > __kmalloc_node_track_caller_noprof(PASS_BUCKET_PARAMS(size, NULL), flags, node, caller) > -#define kmalloc_node_track_caller(...) \ > - alloc_hooks(kmalloc_node_track_caller_noprof(__VA_ARGS__, _RET_IP_)) > +#define kmalloc_node_track_caller(size, ...) \ > + alloc_sized_hooks(kmalloc_node_track_caller_noprof, size, __VA_ARGS__, _RET_IP_) > > /* > * kmalloc_track_caller is a special version of kmalloc that records the > @@ -807,13 +807,13 @@ static inline __alloc_size(1) void *kzalloc_noprof(size_t size, gfp_t flags) > { > return kmalloc_noprof(size, flags | __GFP_ZERO); > } > -#define kzalloc(...) alloc_hooks(kzalloc_noprof(__VA_ARGS__)) > +#define kzalloc(size, ...) alloc_sized_hooks(kzalloc_noprof, size, __VA_ARGS__) > #define kzalloc_node(_size, _flags, _node) kmalloc_node(_size, (_flags)|__GFP_ZERO, _node) > > void *__kvmalloc_node_noprof(DECL_BUCKET_PARAMS(size, b), gfp_t flags, int node) __alloc_size(1); > #define kvmalloc_node_noprof(size, flags, node) \ > __kvmalloc_node_noprof(PASS_BUCKET_PARAMS(size, NULL), flags, node) > -#define kvmalloc_node(...) alloc_hooks(kvmalloc_node_noprof(__VA_ARGS__)) > +#define kvmalloc_node(size, ...) alloc_sized_hooks(kvmalloc_node_noprof, size, __VA_ARGS__) > > #define kvmalloc(_size, _flags) kvmalloc_node(_size, _flags, NUMA_NO_NODE) > #define kvmalloc_noprof(_size, _flags) kvmalloc_node_noprof(_size, _flags, NUMA_NO_NODE) > diff --git a/lib/alloc_tag.c b/lib/alloc_tag.c > index 81e5f9a70f22..6d2cb72bf269 100644 > --- a/lib/alloc_tag.c > +++ b/lib/alloc_tag.c > @@ -78,6 +78,14 @@ static void alloc_tag_to_text(struct seq_buf *out, struct codetag *ct) > > seq_buf_printf(out, "%12lli %8llu ", bytes, counter.calls); > codetag_to_text(out, ct); > +#ifdef CONFIG_SLAB_PER_SITE > + seq_buf_putc(out, ' '); > + seq_buf_printf(out, "size:%s(%zu) slab:%s", > + tag->meta.sized == 0 ? "non-slab" : "non-slab" term sounds overly specific and we might extend this to some other allocations as well in the future. I would suggest "unknown" instead. > + tag->meta.sized == SIZE_MAX ? "dynamic" : "fixed", > + tag->meta.sized == SIZE_MAX ? 0 : tag->meta.sized, > + tag->meta.cache ? "ready" : "unused"); I don't see "struct alloc_meta" having a "cache" member... Since you are changing the format of this file, you want to also bump up the file version inside print_allocinfo_header(). > +#endif > seq_buf_putc(out, ' '); > seq_buf_putc(out, '\n'); > } > diff --git a/mm/Kconfig b/mm/Kconfig > index b72e7d040f78..855c63c3270d 100644 > --- a/mm/Kconfig > +++ b/mm/Kconfig > @@ -296,6 +296,14 @@ config SLAB_BUCKETS > > If unsure, say Y. > > +config SLAB_PER_SITE > + bool "Separate slab allocations by call size" > + depends on !SLUB_TINY > + default SLAB_FREELIST_HARDENED > + select SLAB_BUCKETS > + help > + Track sizes of kmalloc() call sites. > + > config SLUB_STATS > default n > bool "Enable performance statistics" > -- > 2.34.1 >
On Thu, Aug 29, 2024 at 09:00:37AM -0700, Suren Baghdasaryan wrote: > On Fri, Aug 9, 2024 at 12:33 AM Kees Cook <kees@kernel.org> wrote: > [...] > > -#define kmem_cache_alloc(...) alloc_hooks(kmem_cache_alloc_noprof(__VA_ARGS__)) > > +#define kmem_cache_alloc(...) alloc_hooks(kmem_cache_alloc_noprof(__VA_ARGS__)) > > nit: seems like an unnecessary churn. Whoops, yes. This was left over from an earlier pass and I failed to get the whitespace correctly restored. I will fix this this. > > diff --git a/lib/alloc_tag.c b/lib/alloc_tag.c > > index 81e5f9a70f22..6d2cb72bf269 100644 > > --- a/lib/alloc_tag.c > > +++ b/lib/alloc_tag.c > > @@ -78,6 +78,14 @@ static void alloc_tag_to_text(struct seq_buf *out, struct codetag *ct) > > > > seq_buf_printf(out, "%12lli %8llu ", bytes, counter.calls); > > codetag_to_text(out, ct); > > +#ifdef CONFIG_SLAB_PER_SITE > > + seq_buf_putc(out, ' '); > > + seq_buf_printf(out, "size:%s(%zu) slab:%s", > > + tag->meta.sized == 0 ? "non-slab" : > > "non-slab" term sounds overly specific and we might extend this to > some other allocations as well in the future. I would suggest > "unknown" instead. Heh, yeah. I went back and forth on the name for this and went with non-slab because we do know what it isn't. It's not some kind of unexpected state. Maybe "untracked", or "unsized", though both seem inaccurate from certain perspectives. > > > + tag->meta.sized == SIZE_MAX ? "dynamic" : "fixed", > > + tag->meta.sized == SIZE_MAX ? 0 : tag->meta.sized, > > + tag->meta.cache ? "ready" : "unused"); > > I don't see "struct alloc_meta" having a "cache" member... Oops, yes, as you found this should have been associated with the next patch that adds "cache". > Since you are changing the format of this file, you want to also bump > up the file version inside print_allocinfo_header(). Okay, yeah. In that case I'll probably split the report into a separate patch after "cache" is added so there's only a single bump in allocinfo versioning.
diff --git a/include/linux/alloc_tag.h b/include/linux/alloc_tag.h index 8c61ccd161ba..f5d8c5849b82 100644 --- a/include/linux/alloc_tag.h +++ b/include/linux/alloc_tag.h @@ -20,6 +20,19 @@ struct alloc_tag_counters { u64 calls; }; +#ifdef CONFIG_SLAB_PER_SITE +struct alloc_meta { + /* 0 means non-slab, SIZE_MAX means dynamic, and everything else is fixed-size. */ + size_t sized; +}; +#define ALLOC_META_INIT(_size) { \ + .sized = (__builtin_constant_p(_size) ? (_size) : SIZE_MAX), \ + } +#else +struct alloc_meta { }; +#define ALLOC_META_INIT(_size) { } +#endif + /* * An instance of this structure is created in a special ELF section at every * allocation callsite. At runtime, the special section is treated as @@ -27,6 +40,7 @@ struct alloc_tag_counters { */ struct alloc_tag { struct codetag ct; + struct alloc_meta meta; struct alloc_tag_counters __percpu *counters; } __aligned(8); @@ -74,19 +88,21 @@ static inline struct alloc_tag *ct_to_alloc_tag(struct codetag *ct) */ DECLARE_PER_CPU(struct alloc_tag_counters, _shared_alloc_tag); -#define DEFINE_ALLOC_TAG(_alloc_tag) \ +#define DEFINE_ALLOC_TAG(_alloc_tag, _meta_init) \ static struct alloc_tag _alloc_tag __used __aligned(8) \ __section("alloc_tags") = { \ .ct = CODE_TAG_INIT, \ + .meta = _meta_init, \ .counters = &_shared_alloc_tag }; #else /* ARCH_NEEDS_WEAK_PER_CPU */ -#define DEFINE_ALLOC_TAG(_alloc_tag) \ +#define DEFINE_ALLOC_TAG(_alloc_tag, _meta_init) \ static DEFINE_PER_CPU(struct alloc_tag_counters, _alloc_tag_cntr); \ static struct alloc_tag _alloc_tag __used __aligned(8) \ __section("alloc_tags") = { \ .ct = CODE_TAG_INIT, \ + .meta = _meta_init, \ .counters = &_alloc_tag_cntr }; #endif /* ARCH_NEEDS_WEAK_PER_CPU */ @@ -191,7 +207,7 @@ static inline void alloc_tag_sub(union codetag_ref *ref, size_t bytes) #else /* CONFIG_MEM_ALLOC_PROFILING */ -#define DEFINE_ALLOC_TAG(_alloc_tag) +#define DEFINE_ALLOC_TAG(_alloc_tag, _meta_init) static inline bool mem_alloc_profiling_enabled(void) { return false; } static inline void alloc_tag_add(union codetag_ref *ref, struct alloc_tag *tag, size_t bytes) {} @@ -210,8 +226,14 @@ static inline void alloc_tag_sub(union codetag_ref *ref, size_t bytes) {} #define alloc_hooks(_do_alloc) \ ({ \ - DEFINE_ALLOC_TAG(_alloc_tag); \ + DEFINE_ALLOC_TAG(_alloc_tag, { }); \ alloc_hooks_tag(&_alloc_tag, _do_alloc); \ }) +#define alloc_sized_hooks(_do_alloc, _size, ...) \ +({ \ + DEFINE_ALLOC_TAG(_alloc_tag, ALLOC_META_INIT(_size)); \ + alloc_hooks_tag(&_alloc_tag, _do_alloc(_size, __VA_ARGS__)); \ +}) + #endif /* _LINUX_ALLOC_TAG_H */ diff --git a/include/linux/slab.h b/include/linux/slab.h index 86cb61a0102c..314d24c79e05 100644 --- a/include/linux/slab.h +++ b/include/linux/slab.h @@ -541,7 +541,7 @@ static_assert(PAGE_SHIFT <= 20); */ void *kmem_cache_alloc_noprof(struct kmem_cache *cachep, gfp_t flags) __assume_slab_alignment __malloc; -#define kmem_cache_alloc(...) alloc_hooks(kmem_cache_alloc_noprof(__VA_ARGS__)) +#define kmem_cache_alloc(...) alloc_hooks(kmem_cache_alloc_noprof(__VA_ARGS__)) void *kmem_cache_alloc_lru_noprof(struct kmem_cache *s, struct list_lru *lru, gfp_t gfpflags) __assume_slab_alignment __malloc; @@ -685,7 +685,7 @@ static __always_inline __alloc_size(1) void *kmalloc_noprof(size_t size, gfp_t f } return __kmalloc_noprof(size, flags); } -#define kmalloc(...) alloc_hooks(kmalloc_noprof(__VA_ARGS__)) +#define kmalloc(size, ...) alloc_sized_hooks(kmalloc_noprof, size, __VA_ARGS__) #define kmem_buckets_alloc(_b, _size, _flags) \ alloc_hooks(__kmalloc_node_noprof(PASS_BUCKET_PARAMS(_size, _b), _flags, NUMA_NO_NODE)) @@ -708,7 +708,7 @@ static __always_inline __alloc_size(1) void *kmalloc_node_noprof(size_t size, gf } return __kmalloc_node_noprof(PASS_BUCKET_PARAMS(size, NULL), flags, node); } -#define kmalloc_node(...) alloc_hooks(kmalloc_node_noprof(__VA_ARGS__)) +#define kmalloc_node(size, ...) alloc_sized_hooks(kmalloc_node_noprof, size, __VA_ARGS__) /** * kmalloc_array - allocate memory for an array. @@ -726,7 +726,7 @@ static inline __alloc_size(1, 2) void *kmalloc_array_noprof(size_t n, size_t siz return kmalloc_noprof(bytes, flags); return kmalloc_noprof(bytes, flags); } -#define kmalloc_array(...) alloc_hooks(kmalloc_array_noprof(__VA_ARGS__)) +#define kmalloc_array(...) alloc_hooks(kmalloc_array_noprof(__VA_ARGS__)) /** * krealloc_array - reallocate memory for an array. @@ -761,8 +761,8 @@ void *__kmalloc_node_track_caller_noprof(DECL_BUCKET_PARAMS(size, b), gfp_t flag unsigned long caller) __alloc_size(1); #define kmalloc_node_track_caller_noprof(size, flags, node, caller) \ __kmalloc_node_track_caller_noprof(PASS_BUCKET_PARAMS(size, NULL), flags, node, caller) -#define kmalloc_node_track_caller(...) \ - alloc_hooks(kmalloc_node_track_caller_noprof(__VA_ARGS__, _RET_IP_)) +#define kmalloc_node_track_caller(size, ...) \ + alloc_sized_hooks(kmalloc_node_track_caller_noprof, size, __VA_ARGS__, _RET_IP_) /* * kmalloc_track_caller is a special version of kmalloc that records the @@ -807,13 +807,13 @@ static inline __alloc_size(1) void *kzalloc_noprof(size_t size, gfp_t flags) { return kmalloc_noprof(size, flags | __GFP_ZERO); } -#define kzalloc(...) alloc_hooks(kzalloc_noprof(__VA_ARGS__)) +#define kzalloc(size, ...) alloc_sized_hooks(kzalloc_noprof, size, __VA_ARGS__) #define kzalloc_node(_size, _flags, _node) kmalloc_node(_size, (_flags)|__GFP_ZERO, _node) void *__kvmalloc_node_noprof(DECL_BUCKET_PARAMS(size, b), gfp_t flags, int node) __alloc_size(1); #define kvmalloc_node_noprof(size, flags, node) \ __kvmalloc_node_noprof(PASS_BUCKET_PARAMS(size, NULL), flags, node) -#define kvmalloc_node(...) alloc_hooks(kvmalloc_node_noprof(__VA_ARGS__)) +#define kvmalloc_node(size, ...) alloc_sized_hooks(kvmalloc_node_noprof, size, __VA_ARGS__) #define kvmalloc(_size, _flags) kvmalloc_node(_size, _flags, NUMA_NO_NODE) #define kvmalloc_noprof(_size, _flags) kvmalloc_node_noprof(_size, _flags, NUMA_NO_NODE) diff --git a/lib/alloc_tag.c b/lib/alloc_tag.c index 81e5f9a70f22..6d2cb72bf269 100644 --- a/lib/alloc_tag.c +++ b/lib/alloc_tag.c @@ -78,6 +78,14 @@ static void alloc_tag_to_text(struct seq_buf *out, struct codetag *ct) seq_buf_printf(out, "%12lli %8llu ", bytes, counter.calls); codetag_to_text(out, ct); +#ifdef CONFIG_SLAB_PER_SITE + seq_buf_putc(out, ' '); + seq_buf_printf(out, "size:%s(%zu) slab:%s", + tag->meta.sized == 0 ? "non-slab" : + tag->meta.sized == SIZE_MAX ? "dynamic" : "fixed", + tag->meta.sized == SIZE_MAX ? 0 : tag->meta.sized, + tag->meta.cache ? "ready" : "unused"); +#endif seq_buf_putc(out, ' '); seq_buf_putc(out, '\n'); } diff --git a/mm/Kconfig b/mm/Kconfig index b72e7d040f78..855c63c3270d 100644 --- a/mm/Kconfig +++ b/mm/Kconfig @@ -296,6 +296,14 @@ config SLAB_BUCKETS If unsure, say Y. +config SLAB_PER_SITE + bool "Separate slab allocations by call size" + depends on !SLUB_TINY + default SLAB_FREELIST_HARDENED + select SLAB_BUCKETS + help + Track sizes of kmalloc() call sites. + config SLUB_STATS default n bool "Enable performance statistics"
For slab allocations, record whether the call site is using a fixed size (i.e. compile time constant) or a dynamic size. Report the results in /proc/allocinfo. Improvements needed: - examine realloc routines for needed coverage Signed-off-by: Kees Cook <kees@kernel.org> --- Cc: Suren Baghdasaryan <surenb@google.com> Cc: Kent Overstreet <kent.overstreet@linux.dev> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Christoph Lameter <cl@linux.com> Cc: Pekka Enberg <penberg@kernel.org> Cc: David Rientjes <rientjes@google.com> Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com> Cc: Andrew Morton <akpm@linux-foundation.org> Cc: Roman Gushchin <roman.gushchin@linux.dev> Cc: Hyeonggon Yoo <42.hyeyoo@gmail.com> Cc: linux-mm@kvack.org --- include/linux/alloc_tag.h | 30 ++++++++++++++++++++++++++---- include/linux/slab.h | 16 ++++++++-------- lib/alloc_tag.c | 8 ++++++++ mm/Kconfig | 8 ++++++++ 4 files changed, 50 insertions(+), 12 deletions(-)