Message ID | 2a2f08bc8118b369610d34e4d190a879d44f76b8.1739866028.git.maciej.wieczor-retman@intel.com (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | [v2,01/14] kasan: sw_tags: Use arithmetic shift for shadow computation | expand |
On Tue, Feb 18, 2025 at 9:20 AM Maciej Wieczor-Retman <maciej.wieczor-retman@intel.com> wrote: > > On x86, generic KASAN is setup in a way that needs a single > KASAN_SHADOW_OFFSET value for both 4 and 5 level paging. It's required > to facilitate boot time switching and it's a compiler ABI so it can't be > changed during runtime. > > Software tag-based mode doesn't tie shadow start and end to any linear > addresses as part of the compiler ABI so it can be changed during > runtime. KASAN_SHADOW_OFFSET is passed to the compiler via hwasan-mapping-offset, see scripts/Makefile.kasan (for the INLINE mode). So while we can change its value, it has to be known at compile time. So I don't think using a runtime constant would work. Which means that KASAN_SHADOW_OFFSET has to have such a value that works for both 4 and 5 level page tables. This possibly means we might need something different than the first patch in this series. But in case I'm wrong, I left comments for the current code below. > This notion, for KASAN purposes, allows to optimize out macros > such us pgtable_l5_enabled() which would otherwise be used in every > single KASAN related function. > > Use runtime_const infrastructure with pgtable_l5_enabled() to initialize > the end address of KASAN's shadow address space. It's a good choice > since in software tag based mode KASAN_SHADOW_OFFSET and > KASAN_SHADOW_END refer to the same value and the offset in > kasan_mem_to_shadow() is a signed negative value. > > Setup KASAN_SHADOW_END values so that they're aligned to 4TB in 4-level > paging mode and to 2PB in 5-level paging mode. Also update x86 memory > map documentation. > > Signed-off-by: Maciej Wieczor-Retman <maciej.wieczor-retman@intel.com> > --- > Changelog v2: > - Change documentation kasan start address to non-dense values. > > Documentation/arch/x86/x86_64/mm.rst | 6 ++++-- > arch/x86/Kconfig | 3 +-- > arch/x86/include/asm/kasan.h | 14 +++++++++++++- > arch/x86/kernel/vmlinux.lds.S | 1 + > arch/x86/mm/kasan_init_64.c | 5 ++++- > 5 files changed, 23 insertions(+), 6 deletions(-) > > diff --git a/Documentation/arch/x86/x86_64/mm.rst b/Documentation/arch/x86/x86_64/mm.rst > index f2db178b353f..5014ec322e19 100644 > --- a/Documentation/arch/x86/x86_64/mm.rst > +++ b/Documentation/arch/x86/x86_64/mm.rst > @@ -60,7 +60,8 @@ Complete virtual memory map with 4-level page tables > ffffe90000000000 | -23 TB | ffffe9ffffffffff | 1 TB | ... unused hole > ffffea0000000000 | -22 TB | ffffeaffffffffff | 1 TB | virtual memory map (vmemmap_base) > ffffeb0000000000 | -21 TB | ffffebffffffffff | 1 TB | ... unused hole > - ffffec0000000000 | -20 TB | fffffbffffffffff | 16 TB | KASAN shadow memory > + ffffec0000000000 | -20 TB | fffffbffffffffff | 16 TB | KASAN shadow memory (generic mode) > + fffff40000000000 | -8 TB | fffffc0000000000 | 8 TB | KASAN shadow memory (software tag-based mode) > __________________|____________|__________________|_________|____________________________________________________________ > | > | Identical layout to the 56-bit one from here on: > @@ -130,7 +131,8 @@ Complete virtual memory map with 5-level page tables > ffd2000000000000 | -11.5 PB | ffd3ffffffffffff | 0.5 PB | ... unused hole > ffd4000000000000 | -11 PB | ffd5ffffffffffff | 0.5 PB | virtual memory map (vmemmap_base) > ffd6000000000000 | -10.5 PB | ffdeffffffffffff | 2.25 PB | ... unused hole > - ffdf000000000000 | -8.25 PB | fffffbffffffffff | ~8 PB | KASAN shadow memory > + ffdf000000000000 | -8.25 PB | fffffbffffffffff | ~8 PB | KASAN shadow memory (generic mode) > + ffe0000000000000 | -6 PB | fff0000000000000 | 4 PB | KASAN shadow memory (software tag-based mode) > __________________|____________|__________________|_________|____________________________________________________________ > | > | Identical layout to the 47-bit one from here on: > diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig > index 6df7779ed6da..f4ef64bf824a 100644 > --- a/arch/x86/Kconfig > +++ b/arch/x86/Kconfig > @@ -400,8 +400,7 @@ config AUDIT_ARCH > > config KASAN_SHADOW_OFFSET > hex > - depends on KASAN > - default 0xdffffc0000000000 > + default 0xdffffc0000000000 if KASAN_GENERIC Let's put a comment here explaining what happens if !KASAN_GENERIC. Also, as I mentioned in the first patch, we need to figure out what to do with scripts/gdb/linux/kasan.py. > > config HAVE_INTEL_TXT > def_bool y > diff --git a/arch/x86/include/asm/kasan.h b/arch/x86/include/asm/kasan.h > index a75f0748a4b6..4bfd3641af84 100644 > --- a/arch/x86/include/asm/kasan.h > +++ b/arch/x86/include/asm/kasan.h > @@ -5,7 +5,7 @@ > #include <linux/const.h> > #include <linux/kasan-tags.h> > #include <linux/types.h> > -#define KASAN_SHADOW_OFFSET _AC(CONFIG_KASAN_SHADOW_OFFSET, UL) > + > #define KASAN_SHADOW_SCALE_SHIFT 3 > > /* > @@ -14,6 +14,8 @@ > * for kernel really starts from compiler's shadow offset + > * 'kernel address space start' >> KASAN_SHADOW_SCALE_SHIFT > */ > +#ifdef CONFIG_KASAN_GENERIC > +#define KASAN_SHADOW_OFFSET _AC(CONFIG_KASAN_SHADOW_OFFSET, UL) > #define KASAN_SHADOW_START (KASAN_SHADOW_OFFSET + \ > ((-1UL << __VIRTUAL_MASK_SHIFT) >> \ > KASAN_SHADOW_SCALE_SHIFT)) > @@ -24,12 +26,22 @@ > #define KASAN_SHADOW_END (KASAN_SHADOW_START + \ > (1ULL << (__VIRTUAL_MASK_SHIFT - \ > KASAN_SHADOW_SCALE_SHIFT))) > +#endif > + > > #ifndef __ASSEMBLY__ > +#include <asm/runtime-const.h> > #include <linux/bitops.h> > #include <linux/bitfield.h> > #include <linux/bits.h> > > +#ifdef CONFIG_KASAN_SW_TAGS > +extern unsigned long KASAN_SHADOW_END_RC; > +#define KASAN_SHADOW_END runtime_const_ptr(KASAN_SHADOW_END_RC) > +#define KASAN_SHADOW_OFFSET KASAN_SHADOW_END > +#define KASAN_SHADOW_START (KASAN_SHADOW_END - ((UL(1)) << (__VIRTUAL_MASK_SHIFT - KASAN_SHADOW_SCALE_SHIFT))) Any reason these are under __ASSEMBLY__? They seem to belong better together with the CONFIG_KASAN_GENERIC definitions above. > +#endif > + > #define arch_kasan_set_tag(addr, tag) __tag_set(addr, tag) > #define arch_kasan_reset_tag(addr) __tag_reset(addr) > #define arch_kasan_get_tag(addr) __tag_get(addr) > diff --git a/arch/x86/kernel/vmlinux.lds.S b/arch/x86/kernel/vmlinux.lds.S > index 0deb4887d6e9..df6c85f8f48f 100644 > --- a/arch/x86/kernel/vmlinux.lds.S > +++ b/arch/x86/kernel/vmlinux.lds.S > @@ -353,6 +353,7 @@ SECTIONS > > RUNTIME_CONST_VARIABLES > RUNTIME_CONST(ptr, USER_PTR_MAX) > + RUNTIME_CONST(ptr, KASAN_SHADOW_END_RC) > > . = ALIGN(PAGE_SIZE); > > diff --git a/arch/x86/mm/kasan_init_64.c b/arch/x86/mm/kasan_init_64.c > index 299a2144dac4..5ca5862a5cd6 100644 > --- a/arch/x86/mm/kasan_init_64.c > +++ b/arch/x86/mm/kasan_init_64.c > @@ -358,6 +358,9 @@ void __init kasan_init(void) > int i; > > memcpy(early_top_pgt, init_top_pgt, sizeof(early_top_pgt)); > + unsigned long KASAN_SHADOW_END_RC = pgtable_l5_enabled() ? 0xfff0000000000000 : 0xfffffc0000000000; I think defining these constants in arch/x86/include/asm/kasan.h is cleaner than hardcoding them here. > + > + runtime_const_init(ptr, KASAN_SHADOW_END_RC); > > /* > * We use the same shadow offset for 4- and 5-level paging to > @@ -372,7 +375,7 @@ void __init kasan_init(void) > * bunch of things like kernel code, modules, EFI mapping, etc. > * We need to take extra steps to not overwrite them. > */ > - if (pgtable_l5_enabled()) { > + if (pgtable_l5_enabled() && !IS_ENABLED(CONFIG_KASAN_SW_TAGS)) { > void *ptr; > > ptr = (void *)pgd_page_vaddr(*pgd_offset_k(KASAN_SHADOW_END)); > -- > 2.47.1 >
diff --git a/Documentation/arch/x86/x86_64/mm.rst b/Documentation/arch/x86/x86_64/mm.rst index f2db178b353f..5014ec322e19 100644 --- a/Documentation/arch/x86/x86_64/mm.rst +++ b/Documentation/arch/x86/x86_64/mm.rst @@ -60,7 +60,8 @@ Complete virtual memory map with 4-level page tables ffffe90000000000 | -23 TB | ffffe9ffffffffff | 1 TB | ... unused hole ffffea0000000000 | -22 TB | ffffeaffffffffff | 1 TB | virtual memory map (vmemmap_base) ffffeb0000000000 | -21 TB | ffffebffffffffff | 1 TB | ... unused hole - ffffec0000000000 | -20 TB | fffffbffffffffff | 16 TB | KASAN shadow memory + ffffec0000000000 | -20 TB | fffffbffffffffff | 16 TB | KASAN shadow memory (generic mode) + fffff40000000000 | -8 TB | fffffc0000000000 | 8 TB | KASAN shadow memory (software tag-based mode) __________________|____________|__________________|_________|____________________________________________________________ | | Identical layout to the 56-bit one from here on: @@ -130,7 +131,8 @@ Complete virtual memory map with 5-level page tables ffd2000000000000 | -11.5 PB | ffd3ffffffffffff | 0.5 PB | ... unused hole ffd4000000000000 | -11 PB | ffd5ffffffffffff | 0.5 PB | virtual memory map (vmemmap_base) ffd6000000000000 | -10.5 PB | ffdeffffffffffff | 2.25 PB | ... unused hole - ffdf000000000000 | -8.25 PB | fffffbffffffffff | ~8 PB | KASAN shadow memory + ffdf000000000000 | -8.25 PB | fffffbffffffffff | ~8 PB | KASAN shadow memory (generic mode) + ffe0000000000000 | -6 PB | fff0000000000000 | 4 PB | KASAN shadow memory (software tag-based mode) __________________|____________|__________________|_________|____________________________________________________________ | | Identical layout to the 47-bit one from here on: diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig index 6df7779ed6da..f4ef64bf824a 100644 --- a/arch/x86/Kconfig +++ b/arch/x86/Kconfig @@ -400,8 +400,7 @@ config AUDIT_ARCH config KASAN_SHADOW_OFFSET hex - depends on KASAN - default 0xdffffc0000000000 + default 0xdffffc0000000000 if KASAN_GENERIC config HAVE_INTEL_TXT def_bool y diff --git a/arch/x86/include/asm/kasan.h b/arch/x86/include/asm/kasan.h index a75f0748a4b6..4bfd3641af84 100644 --- a/arch/x86/include/asm/kasan.h +++ b/arch/x86/include/asm/kasan.h @@ -5,7 +5,7 @@ #include <linux/const.h> #include <linux/kasan-tags.h> #include <linux/types.h> -#define KASAN_SHADOW_OFFSET _AC(CONFIG_KASAN_SHADOW_OFFSET, UL) + #define KASAN_SHADOW_SCALE_SHIFT 3 /* @@ -14,6 +14,8 @@ * for kernel really starts from compiler's shadow offset + * 'kernel address space start' >> KASAN_SHADOW_SCALE_SHIFT */ +#ifdef CONFIG_KASAN_GENERIC +#define KASAN_SHADOW_OFFSET _AC(CONFIG_KASAN_SHADOW_OFFSET, UL) #define KASAN_SHADOW_START (KASAN_SHADOW_OFFSET + \ ((-1UL << __VIRTUAL_MASK_SHIFT) >> \ KASAN_SHADOW_SCALE_SHIFT)) @@ -24,12 +26,22 @@ #define KASAN_SHADOW_END (KASAN_SHADOW_START + \ (1ULL << (__VIRTUAL_MASK_SHIFT - \ KASAN_SHADOW_SCALE_SHIFT))) +#endif + #ifndef __ASSEMBLY__ +#include <asm/runtime-const.h> #include <linux/bitops.h> #include <linux/bitfield.h> #include <linux/bits.h> +#ifdef CONFIG_KASAN_SW_TAGS +extern unsigned long KASAN_SHADOW_END_RC; +#define KASAN_SHADOW_END runtime_const_ptr(KASAN_SHADOW_END_RC) +#define KASAN_SHADOW_OFFSET KASAN_SHADOW_END +#define KASAN_SHADOW_START (KASAN_SHADOW_END - ((UL(1)) << (__VIRTUAL_MASK_SHIFT - KASAN_SHADOW_SCALE_SHIFT))) +#endif + #define arch_kasan_set_tag(addr, tag) __tag_set(addr, tag) #define arch_kasan_reset_tag(addr) __tag_reset(addr) #define arch_kasan_get_tag(addr) __tag_get(addr) diff --git a/arch/x86/kernel/vmlinux.lds.S b/arch/x86/kernel/vmlinux.lds.S index 0deb4887d6e9..df6c85f8f48f 100644 --- a/arch/x86/kernel/vmlinux.lds.S +++ b/arch/x86/kernel/vmlinux.lds.S @@ -353,6 +353,7 @@ SECTIONS RUNTIME_CONST_VARIABLES RUNTIME_CONST(ptr, USER_PTR_MAX) + RUNTIME_CONST(ptr, KASAN_SHADOW_END_RC) . = ALIGN(PAGE_SIZE); diff --git a/arch/x86/mm/kasan_init_64.c b/arch/x86/mm/kasan_init_64.c index 299a2144dac4..5ca5862a5cd6 100644 --- a/arch/x86/mm/kasan_init_64.c +++ b/arch/x86/mm/kasan_init_64.c @@ -358,6 +358,9 @@ void __init kasan_init(void) int i; memcpy(early_top_pgt, init_top_pgt, sizeof(early_top_pgt)); + unsigned long KASAN_SHADOW_END_RC = pgtable_l5_enabled() ? 0xfff0000000000000 : 0xfffffc0000000000; + + runtime_const_init(ptr, KASAN_SHADOW_END_RC); /* * We use the same shadow offset for 4- and 5-level paging to @@ -372,7 +375,7 @@ void __init kasan_init(void) * bunch of things like kernel code, modules, EFI mapping, etc. * We need to take extra steps to not overwrite them. */ - if (pgtable_l5_enabled()) { + if (pgtable_l5_enabled() && !IS_ENABLED(CONFIG_KASAN_SW_TAGS)) { void *ptr; ptr = (void *)pgd_page_vaddr(*pgd_offset_k(KASAN_SHADOW_END));
On x86, generic KASAN is setup in a way that needs a single KASAN_SHADOW_OFFSET value for both 4 and 5 level paging. It's required to facilitate boot time switching and it's a compiler ABI so it can't be changed during runtime. Software tag-based mode doesn't tie shadow start and end to any linear addresses as part of the compiler ABI so it can be changed during runtime. This notion, for KASAN purposes, allows to optimize out macros such us pgtable_l5_enabled() which would otherwise be used in every single KASAN related function. Use runtime_const infrastructure with pgtable_l5_enabled() to initialize the end address of KASAN's shadow address space. It's a good choice since in software tag based mode KASAN_SHADOW_OFFSET and KASAN_SHADOW_END refer to the same value and the offset in kasan_mem_to_shadow() is a signed negative value. Setup KASAN_SHADOW_END values so that they're aligned to 4TB in 4-level paging mode and to 2PB in 5-level paging mode. Also update x86 memory map documentation. Signed-off-by: Maciej Wieczor-Retman <maciej.wieczor-retman@intel.com> --- Changelog v2: - Change documentation kasan start address to non-dense values. Documentation/arch/x86/x86_64/mm.rst | 6 ++++-- arch/x86/Kconfig | 3 +-- arch/x86/include/asm/kasan.h | 14 +++++++++++++- arch/x86/kernel/vmlinux.lds.S | 1 + arch/x86/mm/kasan_init_64.c | 5 ++++- 5 files changed, 23 insertions(+), 6 deletions(-)