Message ID | 2a9262165c71733792974f5e27795625013bc656.1731069334.git.oleksii.kurochko@gmail.com (mailing list archive) |
---|---|
State | Superseded |
Headers | show |
Series | Setup memory management for RISC-V | expand |
On 08.11.2024 13:51, Oleksii Kurochko wrote: > @@ -37,9 +42,9 @@ static inline void *maddr_to_virt(paddr_t ma) > */ > static inline unsigned long virt_to_maddr(unsigned long va) > { > - if ((va >= DIRECTMAP_VIRT_START) && > + if ((va >= directmap_virt_start) && Is this a valid / necessary change to make? Right now there looks to be nothing immediately below the directmap, yet that would need guaranteeing (e.g. by some BUILD_BIG_ON() or whatever else) if code builds upon that. > (va < (DIRECTMAP_VIRT_START + DIRECTMAP_SIZE))) > - return directmapoff_to_maddr(va - DIRECTMAP_VIRT_START); > + return directmapoff_to_maddr(va - directmap_virt_start); FTAOD - no question about this part of the change. > @@ -423,3 +429,140 @@ void * __init early_fdt_map(paddr_t fdt_paddr) > > return fdt_virt; > } > + > +vaddr_t __ro_after_init directmap_virt_start = DIRECTMAP_VIRT_START; > + > +struct page_info *__ro_after_init frametable_virt_start; As for directmap_virt_start - perhaps better with initializer? > +#ifndef CONFIG_RISCV_32 > + > +/* Map a frame table to cover physical addresses ps through pe */ > +static void __init setup_frametable_mappings(paddr_t ps, paddr_t pe) > +{ > + paddr_t aligned_ps = ROUNDUP(ps, PAGE_SIZE); > + paddr_t aligned_pe = ROUNDDOWN(pe, PAGE_SIZE); > + unsigned long nr_mfns = PFN_DOWN(aligned_pe - aligned_ps); > + unsigned long frametable_size = nr_mfns * sizeof(*frame_table); > + mfn_t base_mfn; > + > + if ( !frametable_virt_start ) > + frametable_virt_start = frame_table - paddr_to_pfn(aligned_ps); If you make this conditional, then you need an "else" (or something that's effectively one) just like you have in setup_directmap_mappings(). Like for the earlier assumption on ps being zero: Assumptions you make on how a function is used want to at least be self-consistent. I.e. here either you assume the function may be called more than once, or you don't. > +static void __init setup_directmap_mappings(unsigned long base_mfn, > + unsigned long nr_mfns) > +{ > + static mfn_t __initdata directmap_mfn_start = INVALID_MFN_INITIALIZER; > + > + unsigned long base_addr = mfn_to_maddr(_mfn(base_mfn)); Seeing this and ... > + unsigned long high_bits_mask = XEN_PT_LEVEL_MAP_MASK(HYP_PT_ROOT_LEVEL); > + > + /* First call sets the directmap physical and virtual offset. */ > + if ( mfn_eq(directmap_mfn_start, INVALID_MFN) ) > + { > + directmap_mfn_start = _mfn(base_mfn); ... this (and more further down) - perhaps better to have the function take mfn_t right away? > + /* > + * The base address may not be aligned to the second level > + * size in case of Sv39 (e.g. 1GB when using 4KB pages). > + * This would prevent superpage mappings for all the regions > + * because the virtual address and machine address should > + * both be suitably aligned. > + * > + * Prevent that by offsetting the start of the directmap virtual > + * address. > + */ > + directmap_virt_start -= > + (base_addr & high_bits_mask) + (base_addr & ~high_bits_mask); Isn't this the same as directmap_virt_start -= base_addr; i.e. no different from what you had a few revisions back? I continue to think that only the low bits matter for the offsetting. > + } > + > + if ( base_mfn < mfn_x(directmap_mfn_start) ) > + panic("can't add directmap mapping at %#lx below directmap start %#lx\n", > + base_mfn, mfn_x(directmap_mfn_start)); > + > + if ( map_pages_to_xen((vaddr_t)mfn_to_virt(base_mfn), > + _mfn(base_mfn), nr_mfns, > + PAGE_HYPERVISOR_RW) ) > + panic("Directmap mappings for [%#"PRIpaddr", %#"PRIpaddr") failed\n", > + mfn_to_maddr(_mfn(base_mfn)), > + mfn_to_maddr(_mfn(base_mfn + nr_mfns))); Maybe worth also logging the error code? > +void __init setup_mm(void) > +{ > + const struct membanks *banks = bootinfo_get_mem(); > + paddr_t ram_start = INVALID_PADDR; > + paddr_t ram_end = 0; > + paddr_t ram_size = 0; > + unsigned int i; > + > + /* > + * We need some memory to allocate the page-tables used for the directmap > + * mappings. But some regions may contain memory already allocated > + * for other uses (e.g. modules, reserved-memory...). > + * > + * For simplicity, add all the free regions in the boot allocator. > + */ > + populate_boot_allocator(); > + > + for ( i = 0; i < banks->nr_banks; i++ ) > + { > + const struct membank *bank = &banks->bank[i]; > + paddr_t bank_start = ROUNDUP(bank->start, PAGE_SIZE); > + paddr_t bank_end = ROUNDDOWN(bank->start + bank->size, PAGE_SIZE); > + unsigned long bank_size = bank_end - bank_start; > + > + ram_size += bank_size; As before - you maintain ram_size here, ... > + ram_start = min(ram_start, bank_start); > + ram_end = max(ram_end, bank_end); > + > + setup_directmap_mappings(PFN_DOWN(bank_start), PFN_DOWN(bank_size)); > + } > + > + setup_frametable_mappings(ram_start, ram_end); > + max_page = PFN_DOWN(ram_end); > +} ... without ever using the value. Why? Jan
On Mon, 2024-11-11 at 11:29 +0100, Jan Beulich wrote: > On 08.11.2024 13:51, Oleksii Kurochko wrote: > > @@ -37,9 +42,9 @@ static inline void *maddr_to_virt(paddr_t ma) > > */ > > static inline unsigned long virt_to_maddr(unsigned long va) > > { > > - if ((va >= DIRECTMAP_VIRT_START) && > > + if ((va >= directmap_virt_start) && > > Is this a valid / necessary change to make? You are right, this not valid change, va value is DIRECTMAP_VIRT_START- relative. > Right now there looks to be > nothing immediately below the directmap, yet that would need > guaranteeing > (e.g. by some BUILD_BIG_ON() or whatever else) if code builds upon > that. It is not really clear how to check that nothing below the directmap is present/used. But IIUC there is no need for this check if properly correct the condition above. > > > (va < (DIRECTMAP_VIRT_START + DIRECTMAP_SIZE))) > > - return directmapoff_to_maddr(va - DIRECTMAP_VIRT_START); > > + return directmapoff_to_maddr(va - directmap_virt_start); > > FTAOD - no question about this part of the change. > > > @@ -423,3 +429,140 @@ void * __init early_fdt_map(paddr_t > > fdt_paddr) > > > > return fdt_virt; > > } > > + > > +vaddr_t __ro_after_init directmap_virt_start = > > DIRECTMAP_VIRT_START; > > + > > +struct page_info *__ro_after_init frametable_virt_start; > > As for directmap_virt_start - perhaps better with initializer? Do you mean to initialized by NULL or frame_table? If to initialize by frame_table then the if-condition won't work properly in setup_frametable_mappings() ( but I think that this condition could be dropped as setup_frametable_mappings() is supposed to be called only once ?! ). And you mentioned about this condition here ... > > > +#ifndef CONFIG_RISCV_32 > > + > > +/* Map a frame table to cover physical addresses ps through pe */ > > +static void __init setup_frametable_mappings(paddr_t ps, paddr_t > > pe) > > +{ > > + paddr_t aligned_ps = ROUNDUP(ps, PAGE_SIZE); > > + paddr_t aligned_pe = ROUNDDOWN(pe, PAGE_SIZE); > > + unsigned long nr_mfns = PFN_DOWN(aligned_pe - aligned_ps); > > + unsigned long frametable_size = nr_mfns * > > sizeof(*frame_table); > > + mfn_t base_mfn; > > + > > + if ( !frametable_virt_start ) > > + frametable_virt_start = frame_table - > > paddr_to_pfn(aligned_ps); > > If you make this conditional, then you need an "else" (or something > that's > effectively one) just like you have in setup_directmap_mappings(). > Like > for the earlier assumption on ps being zero: Assumptions you make on > how > a function is used want to at least be self-consistent. I.e. here > either > you assume the function may be called more than once, or you don't. ... Do we have in Xen something to be sure that the function is called only once or I have to come up with static variable inside the function? > > > +static void __init setup_directmap_mappings(unsigned long > > base_mfn, > > + unsigned long nr_mfns) > > +{ > > + static mfn_t __initdata directmap_mfn_start = > > INVALID_MFN_INITIALIZER; > > + > > + unsigned long base_addr = mfn_to_maddr(_mfn(base_mfn)); > > Seeing this and ... > > > + unsigned long high_bits_mask = > > XEN_PT_LEVEL_MAP_MASK(HYP_PT_ROOT_LEVEL); > > + > > + /* First call sets the directmap physical and virtual offset. > > */ > > + if ( mfn_eq(directmap_mfn_start, INVALID_MFN) ) > > + { > > + directmap_mfn_start = _mfn(base_mfn); > > ... this (and more further down) - perhaps better to have the > function take > mfn_t right away? Agree, it makes sense. I'll update correspondingly. > > > + /* > > + * The base address may not be aligned to the second level > > + * size in case of Sv39 (e.g. 1GB when using 4KB pages). > > + * This would prevent superpage mappings for all the > > regions > > + * because the virtual address and machine address should > > + * both be suitably aligned. > > + * > > + * Prevent that by offsetting the start of the directmap > > virtual > > + * address. > > + */ > > + directmap_virt_start -= > > + (base_addr & high_bits_mask) + (base_addr & > > ~high_bits_mask); > > Isn't this the same as > > directmap_virt_start -= base_addr; > > i.e. no different from what you had a few revisions back? I continue > to > think that only the low bits matter for the offsetting. IIUYC you mean that "(base_addr & ~high_bits_mask)" should be dropped then I agree. Thanks for noticing that. > > > + } > > + > > + if ( base_mfn < mfn_x(directmap_mfn_start) ) > > + panic("can't add directmap mapping at %#lx below directmap > > start %#lx\n", > > + base_mfn, mfn_x(directmap_mfn_start)); > > + > > + if ( map_pages_to_xen((vaddr_t)mfn_to_virt(base_mfn), > > + _mfn(base_mfn), nr_mfns, > > + PAGE_HYPERVISOR_RW) ) > > + panic("Directmap mappings for [%#"PRIpaddr", %#"PRIpaddr") > > failed\n", > > + mfn_to_maddr(_mfn(base_mfn)), > > + mfn_to_maddr(_mfn(base_mfn + nr_mfns))); > > Maybe worth also logging the error code? I am not really understand why do we need that as the use will see what is the issue in the message inside panic(). > > > +void __init setup_mm(void) > > +{ > > + const struct membanks *banks = bootinfo_get_mem(); > > + paddr_t ram_start = INVALID_PADDR; > > + paddr_t ram_end = 0; > > + paddr_t ram_size = 0; > > + unsigned int i; > > + > > + /* > > + * We need some memory to allocate the page-tables used for > > the directmap > > + * mappings. But some regions may contain memory already > > allocated > > + * for other uses (e.g. modules, reserved-memory...). > > + * > > + * For simplicity, add all the free regions in the boot > > allocator. > > + */ > > + populate_boot_allocator(); > > + > > + for ( i = 0; i < banks->nr_banks; i++ ) > > + { > > + const struct membank *bank = &banks->bank[i]; > > + paddr_t bank_start = ROUNDUP(bank->start, PAGE_SIZE); > > + paddr_t bank_end = ROUNDDOWN(bank->start + bank->size, > > PAGE_SIZE); > > + unsigned long bank_size = bank_end - bank_start; > > + > > + ram_size += bank_size; > > As before - you maintain ram_size here, ... > > > + ram_start = min(ram_start, bank_start); > > + ram_end = max(ram_end, bank_end); > > + > > + setup_directmap_mappings(PFN_DOWN(bank_start), > > PFN_DOWN(bank_size)); > > + } > > + > > + setup_frametable_mappings(ram_start, ram_end); > > + max_page = PFN_DOWN(ram_end); > > +} > > ... without ever using the value. Why? I started to use bank_end and bank_size in v2 and ram_size isn't really needed anymore. Thanks. ~ Oleksii
On 11.11.2024 13:51, oleksii.kurochko@gmail.com wrote: > On Mon, 2024-11-11 at 11:29 +0100, Jan Beulich wrote: >> On 08.11.2024 13:51, Oleksii Kurochko wrote: >>> @@ -37,9 +42,9 @@ static inline void *maddr_to_virt(paddr_t ma) >>> */ >>> static inline unsigned long virt_to_maddr(unsigned long va) >>> { >>> - if ((va >= DIRECTMAP_VIRT_START) && >>> + if ((va >= directmap_virt_start) && >> >> Is this a valid / necessary change to make? > You are right, this not valid change, va value is DIRECTMAP_VIRT_START- > relative. > >> Right now there looks to be >> nothing immediately below the directmap, yet that would need >> guaranteeing >> (e.g. by some BUILD_BIG_ON() or whatever else) if code builds upon >> that. > It is not really clear how to check that nothing below the directmap is > present/used. But IIUC there is no need for this check if properly > correct the condition above. Right. >>> (va < (DIRECTMAP_VIRT_START + DIRECTMAP_SIZE))) >>> - return directmapoff_to_maddr(va - DIRECTMAP_VIRT_START); >>> + return directmapoff_to_maddr(va - directmap_virt_start); >> >> FTAOD - no question about this part of the change. >> >>> @@ -423,3 +429,140 @@ void * __init early_fdt_map(paddr_t >>> fdt_paddr) >>> >>> return fdt_virt; >>> } >>> + >>> +vaddr_t __ro_after_init directmap_virt_start = >>> DIRECTMAP_VIRT_START; >>> + >>> +struct page_info *__ro_after_init frametable_virt_start; >> >> As for directmap_virt_start - perhaps better with initializer? > Do you mean to initialized by NULL or frame_table? The latter. > If to initialize by frame_table then the if-condition won't work > properly in setup_frametable_mappings() ( but I think that this > condition could be dropped as setup_frametable_mappings() is supposed > to be called only once ?! ). And you mentioned about this condition > here ... > >> >>> +#ifndef CONFIG_RISCV_32 >>> + >>> +/* Map a frame table to cover physical addresses ps through pe */ >>> +static void __init setup_frametable_mappings(paddr_t ps, paddr_t >>> pe) >>> +{ >>> + paddr_t aligned_ps = ROUNDUP(ps, PAGE_SIZE); >>> + paddr_t aligned_pe = ROUNDDOWN(pe, PAGE_SIZE); >>> + unsigned long nr_mfns = PFN_DOWN(aligned_pe - aligned_ps); >>> + unsigned long frametable_size = nr_mfns * >>> sizeof(*frame_table); >>> + mfn_t base_mfn; >>> + >>> + if ( !frametable_virt_start ) >>> + frametable_virt_start = frame_table - >>> paddr_to_pfn(aligned_ps); >> >> If you make this conditional, then you need an "else" (or something >> that's >> effectively one) just like you have in setup_directmap_mappings(). >> Like >> for the earlier assumption on ps being zero: Assumptions you make on >> how >> a function is used want to at least be self-consistent. I.e. here >> either >> you assume the function may be called more than once, or you don't. > ... > > Do we have in Xen something to be sure that the function is called only > once or I have to come up with static variable inside the function? There's no checking needed. All I'm asking for is that the function either be indeed callable multiple times, or that it not wrongly give the impression that it can be called more than once when it really can't be. >>> + if ( base_mfn < mfn_x(directmap_mfn_start) ) >>> + panic("can't add directmap mapping at %#lx below directmap >>> start %#lx\n", >>> + base_mfn, mfn_x(directmap_mfn_start)); >>> + >>> + if ( map_pages_to_xen((vaddr_t)mfn_to_virt(base_mfn), >>> + _mfn(base_mfn), nr_mfns, >>> + PAGE_HYPERVISOR_RW) ) >>> + panic("Directmap mappings for [%#"PRIpaddr", %#"PRIpaddr") >>> failed\n", >>> + mfn_to_maddr(_mfn(base_mfn)), >>> + mfn_to_maddr(_mfn(base_mfn + nr_mfns))); >> >> Maybe worth also logging the error code? > I am not really understand why do we need that as the use will see what > is the issue in the message inside panic(). If the panic() triggers, the user will see that something went wrong with the given range. The error code may give a hint at _what_ went wrong. Jan
diff --git a/xen/arch/riscv/include/asm/mm.h b/xen/arch/riscv/include/asm/mm.h index ebb142502e..6849092352 100644 --- a/xen/arch/riscv/include/asm/mm.h +++ b/xen/arch/riscv/include/asm/mm.h @@ -12,6 +12,8 @@ #include <asm/page-bits.h> +extern vaddr_t directmap_virt_start; + #define pfn_to_paddr(pfn) ((paddr_t)(pfn) << PAGE_SHIFT) #define paddr_to_pfn(pa) ((unsigned long)((pa) >> PAGE_SHIFT)) @@ -25,8 +27,11 @@ static inline void *maddr_to_virt(paddr_t ma) { - BUG_ON("unimplemented"); - return NULL; + unsigned long va_offset = maddr_to_directmapoff(ma); + + ASSERT(va_offset < DIRECTMAP_SIZE); + + return (void *)(directmap_virt_start + va_offset); } /* @@ -37,9 +42,9 @@ static inline void *maddr_to_virt(paddr_t ma) */ static inline unsigned long virt_to_maddr(unsigned long va) { - if ((va >= DIRECTMAP_VIRT_START) && + if ((va >= directmap_virt_start) && (va < (DIRECTMAP_VIRT_START + DIRECTMAP_SIZE))) - return directmapoff_to_maddr(va - DIRECTMAP_VIRT_START); + return directmapoff_to_maddr(va - directmap_virt_start); BUILD_BUG_ON(XEN_VIRT_SIZE != MB(2)); ASSERT((va >> (PAGETABLE_ORDER + PAGE_SHIFT)) == @@ -127,11 +132,13 @@ struct page_info }; }; +extern struct page_info *frametable_virt_start; + #define frame_table ((struct page_info *)FRAMETABLE_VIRT_START) /* Convert between machine frame numbers and page-info structures. */ -#define mfn_to_page(mfn) (frame_table + mfn_x(mfn)) -#define page_to_mfn(pg) _mfn((pg) - frame_table) +#define mfn_to_page(mfn) (frametable_virt_start + mfn_x(mfn)) +#define page_to_mfn(pg) _mfn((pg) - frametable_virt_start) static inline void *page_to_virt(const struct page_info *pg) { diff --git a/xen/arch/riscv/include/asm/setup.h b/xen/arch/riscv/include/asm/setup.h index c0214a9bf2..844a2f0ef1 100644 --- a/xen/arch/riscv/include/asm/setup.h +++ b/xen/arch/riscv/include/asm/setup.h @@ -5,6 +5,8 @@ #define max_init_domid (0) +void setup_mm(void); + #endif /* ASM__RISCV__SETUP_H */ /* diff --git a/xen/arch/riscv/mm.c b/xen/arch/riscv/mm.c index 27026d803b..988673c313 100644 --- a/xen/arch/riscv/mm.c +++ b/xen/arch/riscv/mm.c @@ -372,6 +372,12 @@ int destroy_xen_mappings(unsigned long s, unsigned long e) return -1; } +void share_xen_page_with_guest(struct page_info *page, struct domain *d, + enum XENSHARE_flags flags) +{ + BUG_ON("unimplemented"); +} + void * __init early_fdt_map(paddr_t fdt_paddr) { /* We are using 2MB superpage for mapping the FDT */ @@ -423,3 +429,140 @@ void * __init early_fdt_map(paddr_t fdt_paddr) return fdt_virt; } + +vaddr_t __ro_after_init directmap_virt_start = DIRECTMAP_VIRT_START; + +struct page_info *__ro_after_init frametable_virt_start; + +#ifndef CONFIG_RISCV_32 + +/* Map a frame table to cover physical addresses ps through pe */ +static void __init setup_frametable_mappings(paddr_t ps, paddr_t pe) +{ + paddr_t aligned_ps = ROUNDUP(ps, PAGE_SIZE); + paddr_t aligned_pe = ROUNDDOWN(pe, PAGE_SIZE); + unsigned long nr_mfns = PFN_DOWN(aligned_pe - aligned_ps); + unsigned long frametable_size = nr_mfns * sizeof(*frame_table); + mfn_t base_mfn; + + if ( !frametable_virt_start ) + frametable_virt_start = frame_table - paddr_to_pfn(aligned_ps); + + if ( frametable_size > FRAMETABLE_SIZE ) + panic("The frametable cannot cover [%#"PRIpaddr", %#"PRIpaddr")\n", + ps, pe); + + /* + * align base_mfn and frametable_size to MB(2) to have superpage mapping + * in map_pages_to_xen() + */ + frametable_size = ROUNDUP(frametable_size, MB(2)); + base_mfn = alloc_boot_pages(frametable_size >> PAGE_SHIFT, PFN_DOWN(MB(2))); + + if ( map_pages_to_xen(FRAMETABLE_VIRT_START, base_mfn, + PFN_DOWN(frametable_size), + PAGE_HYPERVISOR_RW) ) + panic("frametable mappings failed: %#lx -> %#lx\n", + FRAMETABLE_VIRT_START, mfn_x(base_mfn)); + + memset(&frame_table[0], 0, nr_mfns * sizeof(*frame_table)); + memset(&frame_table[nr_mfns], -1, + frametable_size - (nr_mfns * sizeof(*frame_table))); +} + +/* Map the region in the directmap area. */ +static void __init setup_directmap_mappings(unsigned long base_mfn, + unsigned long nr_mfns) +{ + static mfn_t __initdata directmap_mfn_start = INVALID_MFN_INITIALIZER; + + unsigned long base_addr = mfn_to_maddr(_mfn(base_mfn)); + unsigned long high_bits_mask = XEN_PT_LEVEL_MAP_MASK(HYP_PT_ROOT_LEVEL); + + /* First call sets the directmap physical and virtual offset. */ + if ( mfn_eq(directmap_mfn_start, INVALID_MFN) ) + { + directmap_mfn_start = _mfn(base_mfn); + + /* + * The base address may not be aligned to the second level + * size in case of Sv39 (e.g. 1GB when using 4KB pages). + * This would prevent superpage mappings for all the regions + * because the virtual address and machine address should + * both be suitably aligned. + * + * Prevent that by offsetting the start of the directmap virtual + * address. + */ + directmap_virt_start -= + (base_addr & high_bits_mask) + (base_addr & ~high_bits_mask); + } + + if ( base_mfn < mfn_x(directmap_mfn_start) ) + panic("can't add directmap mapping at %#lx below directmap start %#lx\n", + base_mfn, mfn_x(directmap_mfn_start)); + + if ( map_pages_to_xen((vaddr_t)mfn_to_virt(base_mfn), + _mfn(base_mfn), nr_mfns, + PAGE_HYPERVISOR_RW) ) + panic("Directmap mappings for [%#"PRIpaddr", %#"PRIpaddr") failed\n", + mfn_to_maddr(_mfn(base_mfn)), + mfn_to_maddr(_mfn(base_mfn + nr_mfns))); +} + +#else /* CONFIG_RISCV_32 */ +#error setup_{directmap,frametable}_mapping() should be implemented for RV_32 +#endif + +/* + * Setup memory management + * + * RISC-V 64 has a large virtual address space (the minimum supported + * MMU mode is Sv39, which provides GBs of VA space). + * + * The directmap_virt_start is shifted lower in the VA space to + * (DIRECTMAP_VIRT_START - masked_low_bits_of_ram_start_address) to avoid + * wasting a large portion of the directmap space, this also allows for simple + * VA <-> PA translations. Also aligns DIRECTMAP_VIRT_START to a GB boundary + * (for Sv39; for other MMU mode boundaries will be bigger ) by masking the + * higher bits of the RAM start address to enable the use of superpages in + * map_pages_to_xen(). + * + * The frametable is mapped starting from physical address 0, minimizing + * wasted VA space and simplifying page_to_mfn() and mfn_to_page() + * translations. + */ +void __init setup_mm(void) +{ + const struct membanks *banks = bootinfo_get_mem(); + paddr_t ram_start = INVALID_PADDR; + paddr_t ram_end = 0; + paddr_t ram_size = 0; + unsigned int i; + + /* + * We need some memory to allocate the page-tables used for the directmap + * mappings. But some regions may contain memory already allocated + * for other uses (e.g. modules, reserved-memory...). + * + * For simplicity, add all the free regions in the boot allocator. + */ + populate_boot_allocator(); + + for ( i = 0; i < banks->nr_banks; i++ ) + { + const struct membank *bank = &banks->bank[i]; + paddr_t bank_start = ROUNDUP(bank->start, PAGE_SIZE); + paddr_t bank_end = ROUNDDOWN(bank->start + bank->size, PAGE_SIZE); + unsigned long bank_size = bank_end - bank_start; + + ram_size += bank_size; + ram_start = min(ram_start, bank_start); + ram_end = max(ram_end, bank_end); + + setup_directmap_mappings(PFN_DOWN(bank_start), PFN_DOWN(bank_size)); + } + + setup_frametable_mappings(ram_start, ram_end); + max_page = PFN_DOWN(ram_end); +} diff --git a/xen/arch/riscv/setup.c b/xen/arch/riscv/setup.c index e29bd75d7c..2887a18c0c 100644 --- a/xen/arch/riscv/setup.c +++ b/xen/arch/riscv/setup.c @@ -12,6 +12,7 @@ #include <asm/early_printk.h> #include <asm/sbi.h> +#include <asm/setup.h> #include <asm/smp.h> #include <asm/traps.h> @@ -59,6 +60,8 @@ void __init noreturn start_xen(unsigned long bootcpu_id, printk("Command line: %s\n", cmdline); cmdline_parse(cmdline); + setup_mm(); + printk("All set up\n"); machine_halt(); diff --git a/xen/include/xen/macros.h b/xen/include/xen/macros.h index 19caaa8026..cd528fbdb1 100644 --- a/xen/include/xen/macros.h +++ b/xen/include/xen/macros.h @@ -2,6 +2,7 @@ #define __MACROS_H__ #define ROUNDUP(x, a) (((x) + (a) - 1) & ~((a) - 1)) +#define ROUNDDOWN(x, a) ((x) & ~((a) - 1)) #define IS_ALIGNED(val, align) (!((val) & ((align) - 1)))
Introduce the implementation of setup_mm(), which includes: 1. Adding all free regions to the boot allocator, as memory is needed to allocate page tables used for frame table mapping. 2. Calculating RAM size and the RAM end address. 3. Setting up direct map mappings from each RAM bank and initialize directmap_virt_start to keep simple VA <-> PA translation and if RAM_start isn't properly aligned then add an additional alignment to directmap_virt_start to be properly aligned with RAM start to use more superpages to reduce pressure on the TLB. 4. Setting up frame table mappings for range [ram_start, ram_end) and initialize properly frametable_virt_start to have simplified version of mfn_to_page() and page_to_mfn(). 5. Setting up max_page. Update virt_to_maddr() to use introduced directmap_virt_start. Implement maddr_to_virt() function to convert a machine address to a virtual address. This function is specifically designed to be used only for the DIRECTMAP region, so a check has been added to ensure that the address does not exceed DIRECTMAP_SIZE. After the introduction of maddr_to_virt() the following linkage error starts to occur and to avoid it share_xen_page_with_guest() stub is added: riscv64-linux-gnu-ld: prelink.o: in function `tasklet_kill': /build/xen/common/tasklet.c:176: undefined reference to `share_xen_page_with_guest' riscv64-linux-gnu-ld: ./.xen-syms.0: hidden symbol `share_xen_page_with_guest' isn't defined riscv64-linux-gnu-ld: final link failed: bad value Despite the linkger fingering tasklet.c, it's trace.o which has the undefined refenrece: $ find . -name \*.o | while read F; do nm $F | grep share_xen_page_with_guest && echo $F; done U share_xen_page_with_guest ./xen/common/built_in.o U share_xen_page_with_guest ./xen/common/trace.o U share_xen_page_with_guest ./xen/prelink.o Looking at trace.i, there is call of share_xen_page_with_guest() but in case of when maddr_to_virt() is defined as stub ("BUG_ON(); return NULL;") DCE happens and the code is just eliminated. Signed-off-by: Oleksii Kurochko <oleksii.kurochko@gmail.com> --- Change in V4: - s/heap/directmap in log message in setup_directmap_mappings(). - drop local variable total_pages in setup_mm() as it is not used. - call setup_frametable_mappings() for [ram_start,ram_end) range. - setup_frametable_mappings(ps, pe): - update initialization ( validation/invalidation ) of frame table. - add and initialize frameframetable_virt_start variable to cover the case that ps ( RAM start ) could be not eqaul 0 so the proper calculations are needed in mfn_to_page() and page_to_mfn(). - setup_directmap_mapping(): - update the value of directmap_virt_start: add an alignment to the size of HYP_PT_ROOT_LEVEL to have proper alignment so more superpages will be used for mapping. - re-use indirectly directmap_virt_start ( by using mfn_to_virt(base_mfn) ) during mapping of directmap region for RAM bank to not face an issue with possible mapping overlapping during the 2nd invocation of setup_directmap_mapping(). - Update the commit message. --- Changes in V3: - Update the comment the part where DCE should be mentioned and directmap-related things are touched. - Move ROUNDDOWN to <xen/macros.h>. - s/sizeof(struct page_info)/sizeof(*frame_table) in setup_frametable_mapping(). - Updates in setup_frametable_mapping(): - align phys_start (ps) up to a page boundary and align phys_end (pe) down to a page boundary. - Update panic message. - Add the comment about alignment of frametable_size and base_mfn. - invalidate all frame_table entries and then just init with 0 only valid ones. - Double blank lines removing. - Initialize directmap_virt_start with DIRECTMAP_VIRT_START. - Updates in setup_directmap_mapping(): - Drop local variable rc as it is used only once. - Move directmap_mfn_start definition to setup_directmap_mapping() and change __ro_after_init to __initdata. - Update the commit message about alignment of directmap_virt_start. - Move down directmap_virt_start for (base_addr & ~XEN_PT_LEVEL_SIZE(HYP_PT_ROOT_LEVEL)) to not waste a lot of directmap space. - Map DIRECTMAP_VIRT_START + (base_addr & XEN_PT_LEVEL_SIZE(HYP_PT_ROOT_LEVEL)) to _mfn(base_mfn). - Add log of the range in case directmap mapping failure. - Drop XENHEAP_VIRT_START and use directmap_virt_start instead. - Update the comment above setup_mm(). - Update the calculation of bank_start and bank_end in setup_mm() to cover the case when a bank doesn't cover full pages. - Move share_xen_page_with_guest() to riscv/mm.c instead of riscv/stub.c. - Drop inclusion of <xen/pdx.h> in riscv/mm.c as thery is nothing used anymore in riscv/mm.c. - Move variable directmap_virt_start and setup_mm() outside #ifndef CONFIG_RISCV_32 ... #endif as they are likely to be common. --- Changes in V2: - merge patch 2 ( xen/riscv: implement maddr_to_virt() ) to the current one as maddr_to_virt() started to use the thing which are introduced in the current patch. - merge with patch 1 ( xen/riscv: add stub for share_xen_page_with_guest() ) as this linkage issue happens during introduction of maddr_to_virt(). - use mathematical range expressions for log messages. - calculate properly amount of mfns in setup_frametable_mapping() taking into account that ps and pe can be not properly aligned. - drop full stop at the end of debug message. - use PFN_DOWN(framsetable_size) instead of frametable_size >> PAGE_SHIFT. - round down ram_size when it is being accumulated in setup_mm() to guarantee that banks can never have partial pages at their start/end. - call setup_directmap_mappings() only for ram bank regions instead of mapping [0, ram_end] region. - drop directmap_virt_end for now as it isn't used at the moment. - update the commit message. --- xen/arch/riscv/include/asm/mm.h | 19 ++-- xen/arch/riscv/include/asm/setup.h | 2 + xen/arch/riscv/mm.c | 143 +++++++++++++++++++++++++++++ xen/arch/riscv/setup.c | 3 + xen/include/xen/macros.h | 1 + 5 files changed, 162 insertions(+), 6 deletions(-)