Message ID | 93da6ba381604748e2c71e2ebd35e80798ec4bb2.1679934166.git.oleksii.kurochko@gmail.com (mailing list archive) |
---|---|
State | Superseded |
Headers | show |
Series | enable MMU for RISC-V | expand |
On 27.03.2023 19:17, Oleksii Kurochko wrote: > --- a/xen/arch/riscv/include/asm/config.h > +++ b/xen/arch/riscv/include/asm/config.h > @@ -39,12 +39,25 @@ > name: > #endif > > -#define XEN_VIRT_START _AT(UL, 0x80200000) > +#define ADDRESS_SPACE_END (_AC(-1, UL)) > +#define SZ_1G 0x40000000 No need for this - please use GB(1) (see xen/config.h) in its stead. > +#ifdef CONFIG_RISCV_64 > +#define XEN_VIRT_START (ADDRESS_SPACE_END - SZ_1G + 1) I first thought the " + 1" would be rendering the address misaligned. May I suggest (ADDRESS_SPACE_END + 1 - SZ_1G) to help avoiding this impression? > +#else > +#error "RV32 isn't supported"asm Stray "asm" at the end of the line. > --- /dev/null > +++ b/xen/arch/riscv/include/asm/page.h > @@ -0,0 +1,62 @@ > +#ifndef _ASM_RISCV_PAGE_H > +#define _ASM_RISCV_PAGE_H > + > +#include <xen/const.h> > +#include <xen/types.h> > + > +#define VPN_MASK ((unsigned long)(PAGETABLE_ENTRIES - 1)) > + > +#define XEN_PT_LEVEL_ORDER(lvl) ((lvl) * PAGETABLE_ORDER) > +#define XEN_PT_LEVEL_SHIFT(lvl) (XEN_PT_LEVEL_ORDER(lvl) + PAGE_SHIFT) > +#define XEN_PT_LEVEL_SIZE(lvl) (_AT(paddr_t, 1) << XEN_PT_LEVEL_SHIFT(lvl)) > +#define XEN_PT_LEVEL_MAP_MASK(lvl) (~(XEN_PT_LEVEL_SIZE(lvl) - 1)) > +#define XEN_PT_LEVEL_MASK(lvl) (VPN_MASK << XEN_PT_LEVEL_SHIFT(lvl)) > + > +#define PTE_VALID BIT(0, UL) > +#define PTE_READABLE BIT(1, UL) > +#define PTE_WRITABLE BIT(2, UL) > +#define PTE_EXECUTABLE BIT(3, UL) > +#define PTE_USER BIT(4, UL) > +#define PTE_GLOBAL BIT(5, UL) > +#define PTE_ACCESSED BIT(6, UL) > +#define PTE_DIRTY BIT(7, UL) > +#define PTE_RSW (BIT(8, UL) | BIT(9, UL)) > + > +#define PTE_LEAF_DEFAULT (PTE_VALID | PTE_READABLE | PTE_WRITABLE) > +#define PTE_TABLE (PTE_VALID) > + > +/* Calculate the offsets into the pagetables for a given VA */ > +#define pt_linear_offset(lvl, va) ((va) >> XEN_PT_LEVEL_SHIFT(lvl)) > + > +#define pt_index(lvl, va) pt_linear_offset(lvl, (va) & XEN_PT_LEVEL_MASK(lvl)) > + > +#ifdef CONFIG_RISCV_64 > +typedef uint64_t pte_t; > +#else > +typedef uint32_t pte_t; > +#endif Are you sure you don't want to encapsulate these in a struct, for type safety purposes? > +#define addr_to_pte(x) (((x) >> PTE_PPN_SHIFT) << PAGE_SHIFT) This looking like the inverse of ... > +/* Shift the VPN[x] or PPN[x] fields of a virtual or physical address > + * to become the shifted PPN[x] fields of a page table entry */ > +#define addr_to_ppn(x) (((x) >> PAGE_SHIFT) << PTE_PPN_SHIFT) ... this, does that one want to be called "ppn_to_addr()"? Otherwise I can't see how name an operation fit together. > +static inline pte_t paddr_to_pte(const unsigned long paddr, paddr_t? > + const unsigned long permissions) > +{ > + pte_t tmp = (pte_t)addr_to_ppn(paddr); > + return tmp | permissions; > +} > + > +static inline pte_t pte_to_paddr(const unsigned long paddr) A function of this name wants to take pte_t as parameter and return paddr_t (note the type safety remark higher up, as it becomes relevant right here). It also hardly can be implemented ... > +{ > + return (pte_t)addr_to_pte(paddr); ... in terms of a function/macro with effectively the opposite name. > --- /dev/null > +++ b/xen/arch/riscv/mm.c > @@ -0,0 +1,277 @@ > +#include <xen/init.h> > +#include <xen/kernel.h> > + > +#include <asm/early_printk.h> > +#include <asm/config.h> > +#include <asm/csr.h> > +#include <asm/mm.h> > +#include <asm/page.h> > +#include <asm/processor.h> > + > +#define PGTBL_INITIAL_COUNT 8 What does this magic number derive from? (A comment may help.) > +#define PGTBL_ENTRY_AMOUNT (PAGE_SIZE/sizeof(pte_t)) Nit: Style (blanks around binary operators). > +struct mmu_desc { > + unsigned long num_levels; > + uint32_t pgtbl_count; Nit: Style (as before please avoid fixed width types when their use isn't really warranted; afaics unsigned int will do fine here and elsewhere below). > + pte_t *next_pgtbl; > + pte_t *pgtbl_base; > +}; > + > +extern unsigned long __bss_start[]; > +extern unsigned long __bss_end[]; Why are these being added here? They aren't being used afaics. > +extern unsigned char cpu0_boot_stack[STACK_SIZE]; > + > +#define PHYS_OFFSET ((unsigned long)_start - XEN_VIRT_START) > +#define LOAD_TO_LINK(addr) ((addr) - PHYS_OFFSET) > +#define LINK_TO_LOAD(addr) ((addr) + PHYS_OFFSET) > + > +pte_t __section(".bss.page_aligned") __aligned(PAGE_SIZE) > +stage1_pgtbl_root[PGTBL_ENTRY_AMOUNT]; > + > +pte_t __section(".bss.page_aligned") __aligned(PAGE_SIZE) > +stage1_pgtbl_nonroot[PGTBL_INITIAL_COUNT * PGTBL_ENTRY_AMOUNT]; > + > +#define HANDLE_PGTBL() \ > + index = pt_index(curr_lvl_num, page_addr); \ Nit: Style (stray double blanks). > + if ( pte_is_valid(pgtbl[index]) ) { \ Nit: Style (brace placement throughout the macro - in really small macros the way you have it may be okay as an exception, but this macro isn't small). > + /* Find L{ 0-3 } table */ \ > + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); \ > + } else { \ > + /* Allocate new L{0-3} page table */ \ > + if ( mmu_desc->pgtbl_count == PGTBL_INITIAL_COUNT ) { \ > + early_printk("(XEN) No initial table available\n"); \ > + /* panic(), BUG() or ASSERT() aren't ready now. */ \ > + die(); \ > + } \ > + mmu_desc->pgtbl_count++; \ > + pgtbl[index] = paddr_to_pte((unsigned long)mmu_desc->next_pgtbl, \ > + PTE_VALID); \ > + pgtbl = mmu_desc->next_pgtbl; \ > + mmu_desc->next_pgtbl += PGTBL_ENTRY_AMOUNT; \ > + } > + > +static void __init setup_initial_mapping(struct mmu_desc *mmu_desc, > + unsigned long map_start, > + unsigned long map_end, > + unsigned long pa_start, > + unsigned long permission) Nit: Style (indentation, and it feels like I commented on this before). > +{ > + uint32_t index; > + pte_t *pgtbl; > + unsigned long page_addr; > + unsigned int curr_lvl_num; > + pte_t pte_to_be_written; > + unsigned long paddr; > + > + if ( (unsigned long)_start % XEN_PT_LEVEL_SIZE(0) ) { Nit: Style (brace placement). > + early_printk("(XEN) Xen should be loaded at 4k boundary\n"); > + die(); > + } > + > + if ( map_start & ~XEN_PT_LEVEL_MAP_MASK(0) || > + pa_start & ~XEN_PT_LEVEL_MAP_MASK(0) ) { Nit: Style (indentation and brace placement; more below). > + early_printk("(XEN) map and pa start addresses should be aligned\n"); > + /* panic(), BUG() or ASSERT() aren't ready now. */ > + die(); > + } > + > + page_addr = map_start; > + while ( page_addr < map_end ) { > + pgtbl = mmu_desc->pgtbl_base; > + > + switch (mmu_desc->num_levels) > + { > + case 4: /* Level 3 */ > + curr_lvl_num = 3; > + HANDLE_PGTBL(); > + case 3: /* Level 2 */ > + curr_lvl_num = 2; > + HANDLE_PGTBL(); > + case 2: /* Level 1 */ > + curr_lvl_num = 1; > + HANDLE_PGTBL(); Am I getting it right that curr_lvl_num solely exists to pass a parameter to the macro? If so, why doesn't the macro take an appropriate parameter instead? > + case 1: /* Level 0 */ > + index = pt_index(0, page_addr); > + paddr = (page_addr - map_start) + pa_start; > + pte_to_be_written = paddr_to_pte(paddr, permission); > + > + if ( !pte_is_valid(pgtbl[index]) ) > + pgtbl[index] = pte_to_be_written; > + else { > + /* > + * get an adresses of current pte and that one to > + * be written without permission flags > + */ > + unsigned long curr_pte = > + (unsigned long)&pgtbl[index] & (PTE_PPN_SHIFT - 1); > + pte_to_be_written &= (PTE_PPN_SHIFT - 1); > + > + if ( curr_pte != pte_to_be_written ) { > + early_printk("PTE that is intended to write isn't the \ > + same that the once are overwriting\n"); Iirc there are compilers which warn about line continuation characters inside a string literal. Since adjacent string literals are concatenated by the compiler anyway, you'll find elsewhere we have the equivalent of early_printk("PTE that is intended to write isn't the" " same that the once are overwriting\n"); This will also avoid an excessive number of blanks in the middle of the string. > + /* panic(), <asm/bug.h> aren't ready now. */ > + die(); > + } > + } > + } > + > + /* Point to next page */ > + page_addr += XEN_PT_LEVEL_SIZE(0); > + } > +} > + > +static void __init calc_pgtbl_lvls_num(struct mmu_desc *mmu_desc) > +{ > + unsigned long satp_mode = RV_STAGE1_MODE; > + > + /* Number of page table levels */ > + switch (satp_mode) { Nit: Style (missing blanks and brace placement). > + case SATP_MODE_SV32: > + mmu_desc->num_levels = 2; > + break; > + case SATP_MODE_SV39: > + mmu_desc->num_levels = 3; > + break; > + case SATP_MODE_SV48: > + mmu_desc->num_levels = 4; > + break; > + default: > + early_printk("(XEN) Unsupported SATP_MODE\n"); > + while (1); die() (until you have panic()) like above? > + } > +} > + > +static void __init setup_ptes_permission(struct mmu_desc *mmu_desc) > +{ > + pte_t *pgtbl; > + uint32_t index; > + uint32_t permissions = PTE_VALID; > + unsigned long start = LOAD_TO_LINK((unsigned long)_start); > + unsigned long end = LOAD_TO_LINK((unsigned long)_end); > + > + while ( start < end ) > + { > + pgtbl = (pte_t *) mmu_desc->pgtbl_base; Nit: Style (extra blank after closing parenthesis). But afaict you don't need a case here in the first place; as said before, please avoid casts whenever possible. > + switch (mmu_desc->num_levels) Nit: Style (Missing blanks). > + { > + case 4: /* find Level 3 page table*/ > + index = pt_index(3, start); > + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); > + case 3: /* find level 2 page table */ > + index = pt_index(2, start); > + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); > + case 2: /* find level 1 page table */ > + index = pt_index(1, start); > + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); > + case 1: /* find level 0 page table */ > + index = pt_index(0, start); > + > + if ( is_kernel_text(LINK_TO_LOAD(start)) || > + is_kernel_inittext(LINK_TO_LOAD(start)) ) > + permissions |= PTE_EXECUTABLE | PTE_READABLE; > + > + if ( is_kernel_rodata(LINK_TO_LOAD(start)) ) > + permissions |= PTE_READABLE; > + > + pgtbl[index] |= permissions; setup_initial_mapping() has installed R/W mappings. Hence ORing in PTE_READABLE is meaningless here, and neither .text nor .rodata will end up write-protected. The question is ... > + } > + > + start += PAGE_SIZE; > + } > +} > + > +/* > + * setup_initial_pagetables: > + * > + * Build the page tables for Xen that map the following: > + * 1. Calculate page table's level numbers. > + * 2. Init mmu description structure. > + * 3. Check that linker addresses range doesn't overlap > + * with load addresses range > + * 4. Map all linker addresses and load addresses ( it shouldn't > + * be 1:1 mapped and will be 1:1 mapped only in case if > + * linker address is equal to load address ) with > + * RW permissions by default. > + * 5. Setup proper PTE permissions for each section. > + */ > +void __init setup_initial_pagetables(void) > +{ > + struct mmu_desc mmu_desc = { 0, 0, NULL, 0 }; > + > + /* > + * Access to _{stard, end } is always PC-relative > + * thereby when access them we will get load adresses > + * of start and end of Xen > + * To get linker addresses LOAD_TO_LINK() is required > + * to use > + */ > + unsigned long load_start = (unsigned long)_start; > + unsigned long load_end = (unsigned long)_end; > + unsigned long linker_start = LOAD_TO_LINK(load_start); > + unsigned long linker_end = LOAD_TO_LINK(load_end); > + > + if ( (linker_start != load_start) && > + (linker_start <= load_end) && (load_start <= linker_end) ) { > + early_printk("(XEN) linker and load address ranges overlap\n"); > + die(); > + } > + > + calc_pgtbl_lvls_num(&mmu_desc); > + > + mmu_desc.pgtbl_base = stage1_pgtbl_root; > + mmu_desc.next_pgtbl = stage1_pgtbl_nonroot; > + > + setup_initial_mapping(&mmu_desc, > + linker_start, > + linker_end, > + load_start, > + PTE_LEAF_DEFAULT); > + > + setup_ptes_permission(&mmu_desc); ...: Why does this require a 2nd pass / function in the first place? > +} > + > +void __init enable_mmu() > +{ > + /* > + * Calculate a linker time address of the mmu_is_enabled > + * label and update CSR_STVEC with it. > + * MMU is configured in a way where linker addresses are mapped > + * on load addresses so in a case when linker addresses are not equal > + * to load addresses, after MMU is enabled, it will cause > + * an exception and jump to linker time addresses. > + * Otherwise if load addresses are equal to linker addresses the code > + * after mmu_is_enabled label will be executed without exception. > + */ > + csr_write(CSR_STVEC, LOAD_TO_LINK((unsigned long)&&mmu_is_enabled)); That looks like a pretty unusual way of moving from physical to virtual addresses. But if it works ... > + /* Ensure page table writes precede loading the SATP */ > + asm volatile("sfence.vma"); > + > + /* Enable the MMU and load the new pagetable for Xen */ > + csr_write(CSR_SATP, > + ((unsigned long)stage1_pgtbl_root >> PAGE_SHIFT) | RV_STAGE1_MODE << SATP_MODE_SHIFT); Nit: Style (excessively long line). > + asm volatile(".align 2"); > +mmu_is_enabled: Nit: Style (labels want indenting by at least one blank). > @@ -26,3 +27,13 @@ void __init noreturn start_xen(unsigned long bootcpu_id, > > unreachable(); > } > + > +void __init noreturn cont_after_mmu_is_enabled(void) To prevent undue link-time optimization such a function also wants to be noinline. Jan
Hi, On 28/03/2023 16:34, Jan Beulich wrote: > On 27.03.2023 19:17, Oleksii Kurochko wrote: >> --- a/xen/arch/riscv/include/asm/config.h >> +++ b/xen/arch/riscv/include/asm/config.h >> @@ -39,12 +39,25 @@ >> name: >> #endif >> >> -#define XEN_VIRT_START _AT(UL, 0x80200000) >> +#define ADDRESS_SPACE_END (_AC(-1, UL)) >> +#define SZ_1G 0x40000000 > > No need for this - please use GB(1) (see xen/config.h) in its stead. > >> +#ifdef CONFIG_RISCV_64 >> +#define XEN_VIRT_START (ADDRESS_SPACE_END - SZ_1G + 1) > > I first thought the " + 1" would be rendering the address misaligned. > May I suggest (ADDRESS_SPACE_END + 1 - SZ_1G) to help avoiding this > impression? I will jump on this to say that I am finding quite difficult to review code using define/variable that contains "end" in their name because it is never clear whether this is inclusive or exclusive. In this case, it is inclusive, but within the same patch I can see map_end which is exclusive. For this case, my suggestion would be to remove ADDRESS_SPACE_END and just hardcode the value where you want to put Xen. For others, you could use (start, size) to describe a region. Also, are you sure that all the CPU will be able to support more full 64-bit VA space? Cheers,
On Tue, 2023-03-28 at 17:34 +0200, Jan Beulich wrote: > On 27.03.2023 19:17, Oleksii Kurochko wrote: > > --- a/xen/arch/riscv/include/asm/config.h > > +++ b/xen/arch/riscv/include/asm/config.h > > @@ -39,12 +39,25 @@ > > name: > > #endif > > > > -#define XEN_VIRT_START _AT(UL, 0x80200000) > > +#define ADDRESS_SPACE_END (_AC(-1, UL)) > > +#define SZ_1G 0x40000000 > > No need for this - please use GB(1) (see xen/config.h) in its stead. I'll use GB(1). Thanks. > > > +#ifdef CONFIG_RISCV_64 > > +#define XEN_VIRT_START (ADDRESS_SPACE_END - SZ_1G + 1) > > I first thought the " + 1" would be rendering the address misaligned. > May I suggest (ADDRESS_SPACE_END + 1 - SZ_1G) to help avoiding this > impression? Sure. Will update it in next version of the patch series. > > > +#else > > +#error "RV32 isn't supported"asm > > Stray "asm" at the end of the line. My fault... > > > --- /dev/null > > +++ b/xen/arch/riscv/include/asm/page.h > > @@ -0,0 +1,62 @@ > > +#ifndef _ASM_RISCV_PAGE_H > > +#define _ASM_RISCV_PAGE_H > > + > > +#include <xen/const.h> > > +#include <xen/types.h> > > + > > +#define VPN_MASK ((unsigned > > long)(PAGETABLE_ENTRIES - 1)) > > + > > +#define XEN_PT_LEVEL_ORDER(lvl) ((lvl) * PAGETABLE_ORDER) > > +#define XEN_PT_LEVEL_SHIFT(lvl) (XEN_PT_LEVEL_ORDER(lvl) + > > PAGE_SHIFT) > > +#define XEN_PT_LEVEL_SIZE(lvl) (_AT(paddr_t, 1) << > > XEN_PT_LEVEL_SHIFT(lvl)) > > +#define XEN_PT_LEVEL_MAP_MASK(lvl) (~(XEN_PT_LEVEL_SIZE(lvl) - > > 1)) > > +#define XEN_PT_LEVEL_MASK(lvl) (VPN_MASK << > > XEN_PT_LEVEL_SHIFT(lvl)) > > + > > +#define PTE_VALID BIT(0, UL) > > +#define PTE_READABLE BIT(1, UL) > > +#define PTE_WRITABLE BIT(2, UL) > > +#define PTE_EXECUTABLE BIT(3, UL) > > +#define PTE_USER BIT(4, UL) > > +#define PTE_GLOBAL BIT(5, UL) > > +#define PTE_ACCESSED BIT(6, UL) > > +#define PTE_DIRTY BIT(7, UL) > > +#define PTE_RSW (BIT(8, UL) | BIT(9, UL)) > > + > > +#define PTE_LEAF_DEFAULT (PTE_VALID | PTE_READABLE | > > PTE_WRITABLE) > > +#define PTE_TABLE (PTE_VALID) > > + > > +/* Calculate the offsets into the pagetables for a given VA */ > > +#define pt_linear_offset(lvl, va) ((va) >> > > XEN_PT_LEVEL_SHIFT(lvl)) > > + > > +#define pt_index(lvl, va) pt_linear_offset(lvl, (va) & > > XEN_PT_LEVEL_MASK(lvl)) > > + > > +#ifdef CONFIG_RISCV_64 > > +typedef uint64_t pte_t; > > +#else > > +typedef uint32_t pte_t; > > +#endif > > Are you sure you don't want to encapsulate these in a struct, for > type > safety purposes? Yeah, it makes sense. > > > +#define addr_to_pte(x) (((x) >> PTE_PPN_SHIFT) << PAGE_SHIFT) > > This looking like the inverse of ... > > > +/* Shift the VPN[x] or PPN[x] fields of a virtual or physical > > address > > + * to become the shifted PPN[x] fields of a page table entry */ > > +#define addr_to_ppn(x) (((x) >> PAGE_SHIFT) << PTE_PPN_SHIFT) > > ... this, does that one want to be called "ppn_to_addr()"? Otherwise > I > can't see how name an operation fit together. ppn_to_addr() will be better so it will be update in next version of the patch series. > > > +static inline pte_t paddr_to_pte(const unsigned long paddr, > > paddr_t? It looks like I used paddr_t before but got a recommendation to change it to 'ul'. I will double-check. > > > + const unsigned long permissions) > > +{ > > + pte_t tmp = (pte_t)addr_to_ppn(paddr); > > + return tmp | permissions; > > +} > > + > > +static inline pte_t pte_to_paddr(const unsigned long paddr) > > A function of this name wants to take pte_t as parameter and return > paddr_t (note the type safety remark higher up, as it becomes > relevant > right here). It also hardly can be implemented ... > > > +{ > > + return (pte_t)addr_to_pte(paddr); > > ... in terms of a function/macro with effectively the opposite name. Thanks. I'll update that. > > > --- /dev/null > > +++ b/xen/arch/riscv/mm.c > > @@ -0,0 +1,277 @@ > > +#include <xen/init.h> > > +#include <xen/kernel.h> > > + > > +#include <asm/early_printk.h> > > +#include <asm/config.h> > > +#include <asm/csr.h> > > +#include <asm/mm.h> > > +#include <asm/page.h> > > +#include <asm/processor.h> > > + > > +#define PGTBL_INITIAL_COUNT 8 > > What does this magic number derive from? (A comment may help.) It can be now 4 tables as it is enough to map 2Mb. 8 it was when I experimented with bigger sizes of Xen binary and verified that logic of working with page tables works. > > > +#define PGTBL_ENTRY_AMOUNT (PAGE_SIZE/sizeof(pte_t)) > > Nit: Style (blanks around binary operators). Thanks. > > > +struct mmu_desc { > > + unsigned long num_levels; > > + uint32_t pgtbl_count; > > Nit: Style (as before please avoid fixed width types when their use > isn't really warranted; afaics unsigned int will do fine here and > elsewhere below). I will change it but I am not sure that I fully understand what is an issue with uint32_t. > > > + pte_t *next_pgtbl; > > + pte_t *pgtbl_base; > > +}; > > + > > +extern unsigned long __bss_start[]; > > +extern unsigned long __bss_end[]; > > Why are these being added here? They aren't being used afaics. Misssed to remove them. They was used in the previous version of the patch. > > > +extern unsigned char cpu0_boot_stack[STACK_SIZE]; > > + > > +#define PHYS_OFFSET ((unsigned long)_start - XEN_VIRT_START) > > +#define LOAD_TO_LINK(addr) ((addr) - PHYS_OFFSET) > > +#define LINK_TO_LOAD(addr) ((addr) + PHYS_OFFSET) > > + > > +pte_t __section(".bss.page_aligned") __aligned(PAGE_SIZE) > > +stage1_pgtbl_root[PGTBL_ENTRY_AMOUNT]; > > + > > +pte_t __section(".bss.page_aligned") __aligned(PAGE_SIZE) > > +stage1_pgtbl_nonroot[PGTBL_INITIAL_COUNT * PGTBL_ENTRY_AMOUNT]; > > + > > +#define > > HANDLE_PGTBL() > > \ > > + index = pt_index(curr_lvl_num, > > page_addr); \ > > Nit: Style (stray double blanks). Sure. Thanks. > > > + if ( pte_is_valid(pgtbl[index]) ) > > { \ > > Nit: Style (brace placement throughout the macro - in really small > macros the way you have it may be okay as an exception, but this > macro > isn't small). It looks like I confused brace placement for if/else with LK... Anyway it will be updated. > > > + /* Find L{ 0-3 } table > > */ \ > > + pgtbl = (pte_t > > *)pte_to_paddr(pgtbl[index]); \ > > + } else > > { \ > > + /* Allocate new L{0-3} page table > > */ \ > > + if ( mmu_desc->pgtbl_count == PGTBL_INITIAL_COUNT ) > > { \ > > + early_printk("(XEN) No initial table > > available\n"); \ > > + /* panic(), BUG() or ASSERT() aren't ready now. > > */ \ > > + > > die(); \ > > + > > } > > \ > > + mmu_desc- > > >pgtbl_count++; \ > > + pgtbl[index] = paddr_to_pte((unsigned long)mmu_desc- > > >next_pgtbl, \ > > + > > PTE_VALID); \ > > + pgtbl = mmu_desc- > > >next_pgtbl; \ > > + mmu_desc->next_pgtbl += > > PGTBL_ENTRY_AMOUNT; \ > > + } > > + > > +static void __init setup_initial_mapping(struct mmu_desc > > *mmu_desc, > > + unsigned long map_start, > > + unsigned long map_end, > > + unsigned long pa_start, > > + unsigned long permission) > > Nit: Style (indentation, and it feels like I commented on this > before). You commented. I will double check because I though that I fixed it. > > > +{ > > + uint32_t index; > > + pte_t *pgtbl; > > + unsigned long page_addr; > > + unsigned int curr_lvl_num; > > + pte_t pte_to_be_written; > > + unsigned long paddr; > > + > > + if ( (unsigned long)_start % XEN_PT_LEVEL_SIZE(0) ) { > > Nit: Style (brace placement). Thanks. Ill update brace placement. > > > + early_printk("(XEN) Xen should be loaded at 4k > > boundary\n"); > > + die(); > > + } > > + > > + if ( map_start & ~XEN_PT_LEVEL_MAP_MASK(0) || > > + pa_start & ~XEN_PT_LEVEL_MAP_MASK(0) ) { > > Nit: Style (indentation and brace placement; more below). Thanks. Ill update brace placement. > > > + early_printk("(XEN) map and pa start addresses should be > > aligned\n"); > > + /* panic(), BUG() or ASSERT() aren't ready now. */ > > + die(); > > + } > > + > > + page_addr = map_start; > > + while ( page_addr < map_end ) { > > + pgtbl = mmu_desc->pgtbl_base; > > + > > + switch (mmu_desc->num_levels) > > + { > > + case 4: /* Level 3 */ > > + curr_lvl_num = 3; > > + HANDLE_PGTBL(); > > + case 3: /* Level 2 */ > > + curr_lvl_num = 2; > > + HANDLE_PGTBL(); > > + case 2: /* Level 1 */ > > + curr_lvl_num = 1; > > + HANDLE_PGTBL(); > > Am I getting it right that curr_lvl_num solely exists to pass a > parameter > to the macro? If so, why doesn't the macro take an appropriate > parameter > instead? curr_lvl_num exists only to pass current page table level to the macro. I'll add a parameter to the macro. Thanks. > > > + case 1: /* Level 0 */ > > + index = pt_index(0, page_addr); > > + paddr = (page_addr - map_start) + pa_start; > > + pte_to_be_written = paddr_to_pte(paddr, permission); > > + > > + if ( !pte_is_valid(pgtbl[index]) ) > > + pgtbl[index] = pte_to_be_written; > > + else { > > + /* > > + * get an adresses of current pte and that one to > > + * be written without permission flags > > + */ > > + unsigned long curr_pte = > > + (unsigned long)&pgtbl[index] & (PTE_PPN_SHIFT > > - 1); > > + pte_to_be_written &= (PTE_PPN_SHIFT - 1); > > + > > + if ( curr_pte != pte_to_be_written ) { > > + early_printk("PTE that is intended to write > > isn't the \ > > + same that the once are > > overwriting\n"); > > Iirc there are compilers which warn about line continuation > characters > inside a string literal. Since adjacent string literals are > concatenated > by the compiler anyway, you'll find elsewhere we have the equivalent > of > > early_printk("PTE that is intended to write isn't > the" > " same that the once are > overwriting\n"); > > This will also avoid an excessive number of blanks in the middle of > the > string. Agree. It will be much better. > > > + /* panic(), <asm/bug.h> aren't ready now. */ > > + die(); > > + } > > + } > > + } > > + > > + /* Point to next page */ > > + page_addr += XEN_PT_LEVEL_SIZE(0); > > + } > > +} > > + > > +static void __init calc_pgtbl_lvls_num(struct mmu_desc *mmu_desc) > > +{ > > + unsigned long satp_mode = RV_STAGE1_MODE; > > + > > + /* Number of page table levels */ > > + switch (satp_mode) { > > Nit: Style (missing blanks and brace placement). I think that after this patch I'll remmeber about brace placement. Thank you for remind me that :) > > > + case SATP_MODE_SV32: > > + mmu_desc->num_levels = 2; > > + break; > > + case SATP_MODE_SV39: > > + mmu_desc->num_levels = 3; > > + break; > > + case SATP_MODE_SV48: > > + mmu_desc->num_levels = 4; > > + break; > > + default: > > + early_printk("(XEN) Unsupported SATP_MODE\n"); > > + while (1); > > die() (until you have panic()) like above? Should be die. Thanks. > > > + } > > +} > > + > > +static void __init setup_ptes_permission(struct mmu_desc > > *mmu_desc) > > +{ > > + pte_t *pgtbl; > > + uint32_t index; > > + uint32_t permissions = PTE_VALID; > > + unsigned long start = LOAD_TO_LINK((unsigned long)_start); > > + unsigned long end = LOAD_TO_LINK((unsigned long)_end); > > + > > + while ( start < end ) > > + { > > + pgtbl = (pte_t *) mmu_desc->pgtbl_base; > > Nit: Style (extra blank after closing parenthesis). But afaict you > don't > need a case here in the first place; as said before, please avoid > casts > whenever possible. There is no any sense in cast. Thanks. > > > + switch (mmu_desc->num_levels) > > Nit: Style (Missing blanks). > > > + { > > + case 4: /* find Level 3 page table*/ > > + index = pt_index(3, start); > > + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); > > + case 3: /* find level 2 page table */ > > + index = pt_index(2, start); > > + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); > > + case 2: /* find level 1 page table */ > > + index = pt_index(1, start); > > + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); > > + case 1: /* find level 0 page table */ > > + index = pt_index(0, start); > > + > > + if ( is_kernel_text(LINK_TO_LOAD(start)) || > > + is_kernel_inittext(LINK_TO_LOAD(start)) ) > > + permissions |= PTE_EXECUTABLE | PTE_READABLE; > > + > > + if ( is_kernel_rodata(LINK_TO_LOAD(start)) ) > > + permissions |= PTE_READABLE; > > + > > + pgtbl[index] |= permissions; > > setup_initial_mapping() has installed R/W mappings. Hence ORing in > PTE_READABLE is meaningless here, and neither .text nor .rodata > will end up write-protected. The question is ... > > > + } > > + > > + start += PAGE_SIZE; > > + } > > +} > > + > > +/* > > + * setup_initial_pagetables: > > + * > > + * Build the page tables for Xen that map the following: > > + * 1. Calculate page table's level numbers. > > + * 2. Init mmu description structure. > > + * 3. Check that linker addresses range doesn't overlap > > + * with load addresses range > > + * 4. Map all linker addresses and load addresses ( it shouldn't > > + * be 1:1 mapped and will be 1:1 mapped only in case if > > + * linker address is equal to load address ) with > > + * RW permissions by default. > > + * 5. Setup proper PTE permissions for each section. > > + */ > > +void __init setup_initial_pagetables(void) > > +{ > > + struct mmu_desc mmu_desc = { 0, 0, NULL, 0 }; > > + > > + /* > > + * Access to _{stard, end } is always PC-relative > > + * thereby when access them we will get load adresses > > + * of start and end of Xen > > + * To get linker addresses LOAD_TO_LINK() is required > > + * to use > > + */ > > + unsigned long load_start = (unsigned long)_start; > > + unsigned long load_end = (unsigned long)_end; > > + unsigned long linker_start = LOAD_TO_LINK(load_start); > > + unsigned long linker_end = LOAD_TO_LINK(load_end); > > + > > + if ( (linker_start != load_start) && > > + (linker_start <= load_end) && (load_start <= linker_end) > > ) { > > + early_printk("(XEN) linker and load address ranges > > overlap\n"); > > + die(); > > + } > > + > > + calc_pgtbl_lvls_num(&mmu_desc); > > + > > + mmu_desc.pgtbl_base = stage1_pgtbl_root; > > + mmu_desc.next_pgtbl = stage1_pgtbl_nonroot; > > + > > + setup_initial_mapping(&mmu_desc, > > + linker_start, > > + linker_end, > > + load_start, > > + PTE_LEAF_DEFAULT); > > + > > + setup_ptes_permission(&mmu_desc); > > ...: Why does this require a 2nd pass / function in the first place? Probably I misunderstood Julien and it setup_pte_permission can be done in setup_initial_mapping() but here is the reply: https://lore.kernel.org/xen-devel/79e83610-5980-d9b5-7994-6b0cb2b9049a@xen.org/ > > > +} > > + > > +void __init enable_mmu() > > +{ > > + /* > > + * Calculate a linker time address of the mmu_is_enabled > > + * label and update CSR_STVEC with it. > > + * MMU is configured in a way where linker addresses are > > mapped > > + * on load addresses so in a case when linker addresses are > > not equal > > + * to load addresses, after MMU is enabled, it will cause > > + * an exception and jump to linker time addresses. > > + * Otherwise if load addresses are equal to linker addresses > > the code > > + * after mmu_is_enabled label will be executed without > > exception. > > + */ > > + csr_write(CSR_STVEC, LOAD_TO_LINK((unsigned > > long)&&mmu_is_enabled)); > > That looks like a pretty unusual way of moving from physical to > virtual > addresses. But if it works ... > > > + /* Ensure page table writes precede loading the SATP */ > > + asm volatile("sfence.vma"); > > + > > + /* Enable the MMU and load the new pagetable for Xen */ > > + csr_write(CSR_SATP, > > + ((unsigned long)stage1_pgtbl_root >> PAGE_SHIFT) | > > RV_STAGE1_MODE << SATP_MODE_SHIFT); > > Nit: Style (excessively long line). Sure. I will update it. > > > + asm volatile(".align 2"); > > +mmu_is_enabled: > > Nit: Style (labels want indenting by at least one blank). Thanks. I will take it into account. > > > @@ -26,3 +27,13 @@ void __init noreturn start_xen(unsigned long > > bootcpu_id, > > > > unreachable(); > > } > > + > > +void __init noreturn cont_after_mmu_is_enabled(void) > > To prevent undue link-time optimization such a function also wants to > be noinline. Sure. Missed that. ~ Oleksii
Hi Julien, On Tue, 2023-03-28 at 16:44 +0100, Julien Grall wrote: > Hi, > > On 28/03/2023 16:34, Jan Beulich wrote: > > On 27.03.2023 19:17, Oleksii Kurochko wrote: > > > --- a/xen/arch/riscv/include/asm/config.h > > > +++ b/xen/arch/riscv/include/asm/config.h > > > @@ -39,12 +39,25 @@ > > > name: > > > #endif > > > > > > -#define XEN_VIRT_START _AT(UL, 0x80200000) > > > +#define ADDRESS_SPACE_END (_AC(-1, UL)) > > > +#define SZ_1G 0x40000000 > > > > No need for this - please use GB(1) (see xen/config.h) in its > > stead. > > > > > +#ifdef CONFIG_RISCV_64 > > > +#define XEN_VIRT_START (ADDRESS_SPACE_END - SZ_1G + 1) > > > > I first thought the " + 1" would be rendering the address > > misaligned. > > May I suggest (ADDRESS_SPACE_END + 1 - SZ_1G) to help avoiding this > > impression? > > I will jump on this to say that I am finding quite difficult to > review > code using define/variable that contains "end" in their name because > it > is never clear whether this is inclusive or exclusive. > > In this case, it is inclusive, but within the same patch I can see > map_end which is exclusive. > > For this case, my suggestion would be to remove ADDRESS_SPACE_END and > just hardcode the value where you want to put Xen. For others, you > could > use (start, size) to describe a region. Thanks for the suggestion. I'll take it into account. > > Also, are you sure that all the CPU will be able to support more full > 64-bit VA space? I am not sure but based on Linux it looks like it is true. ( at least, they use the same definitions for RV64 ). I am curious how that can be checked. ~ Oleksii
On 29.03.2023 13:43, Oleksii wrote: > On Tue, 2023-03-28 at 17:34 +0200, Jan Beulich wrote: >> On 27.03.2023 19:17, Oleksii Kurochko wrote: >>> --- /dev/null >>> +++ b/xen/arch/riscv/mm.c >>> @@ -0,0 +1,277 @@ >>> +#include <xen/init.h> >>> +#include <xen/kernel.h> >>> + >>> +#include <asm/early_printk.h> >>> +#include <asm/config.h> >>> +#include <asm/csr.h> >>> +#include <asm/mm.h> >>> +#include <asm/page.h> >>> +#include <asm/processor.h> >>> + >>> +#define PGTBL_INITIAL_COUNT 8 >> >> What does this magic number derive from? (A comment may help.) > It can be now 4 tables as it is enough to map 2Mb. 8 it was when I > experimented with bigger sizes of Xen binary and verified that logic of > working with page tables works. Since this is connected to Xen's image size as you say, I guess the constant wants to move to a header, and then be used in an assertion in xen.lds.S. That way one will become easily aware if/when this 2Mb assumption breaks. >>> +struct mmu_desc { >>> + unsigned long num_levels; >>> + uint32_t pgtbl_count; >> >> Nit: Style (as before please avoid fixed width types when their use >> isn't really warranted; afaics unsigned int will do fine here and >> elsewhere below). > I will change it but I am not sure that I fully understand what is an > issue with uint32_t. The issue is simply that ./CODING_STYLE says to prefer basic types whenever possible. >>> +void __init setup_initial_pagetables(void) >>> +{ >>> + struct mmu_desc mmu_desc = { 0, 0, NULL, 0 }; >>> + >>> + /* >>> + * Access to _{stard, end } is always PC-relative >>> + * thereby when access them we will get load adresses >>> + * of start and end of Xen >>> + * To get linker addresses LOAD_TO_LINK() is required >>> + * to use >>> + */ >>> + unsigned long load_start = (unsigned long)_start; >>> + unsigned long load_end = (unsigned long)_end; >>> + unsigned long linker_start = LOAD_TO_LINK(load_start); >>> + unsigned long linker_end = LOAD_TO_LINK(load_end); >>> + >>> + if ( (linker_start != load_start) && >>> + (linker_start <= load_end) && (load_start <= linker_end) >>> ) { >>> + early_printk("(XEN) linker and load address ranges >>> overlap\n"); >>> + die(); >>> + } >>> + >>> + calc_pgtbl_lvls_num(&mmu_desc); >>> + >>> + mmu_desc.pgtbl_base = stage1_pgtbl_root; >>> + mmu_desc.next_pgtbl = stage1_pgtbl_nonroot; >>> + >>> + setup_initial_mapping(&mmu_desc, >>> + linker_start, >>> + linker_end, >>> + load_start, >>> + PTE_LEAF_DEFAULT); >>> + >>> + setup_ptes_permission(&mmu_desc); >> >> ...: Why does this require a 2nd pass / function in the first place? > Probably I misunderstood Julien and it setup_pte_permission can be done > in setup_initial_mapping() but here is the reply: > https://lore.kernel.org/xen-devel/79e83610-5980-d9b5-7994-6b0cb2b9049a@xen.org/ Hmm, yes, his option 2 looks like what you've implemented. Still I don't see why the permissions can't be got right on the first pass. Jan
On 29.03.2023 13:47, Oleksii wrote: > On Tue, 2023-03-28 at 16:44 +0100, Julien Grall wrote: >> On 28/03/2023 16:34, Jan Beulich wrote: >>> On 27.03.2023 19:17, Oleksii Kurochko wrote: >>>> --- a/xen/arch/riscv/include/asm/config.h >>>> +++ b/xen/arch/riscv/include/asm/config.h >>>> @@ -39,12 +39,25 @@ >>>> name: >>>> #endif >>>> >>>> -#define XEN_VIRT_START _AT(UL, 0x80200000) >>>> +#define ADDRESS_SPACE_END (_AC(-1, UL)) >>>> +#define SZ_1G 0x40000000 >>> >>> No need for this - please use GB(1) (see xen/config.h) in its >>> stead. >>> >>>> +#ifdef CONFIG_RISCV_64 >>>> +#define XEN_VIRT_START (ADDRESS_SPACE_END - SZ_1G + 1) >>> >>> I first thought the " + 1" would be rendering the address >>> misaligned. >>> May I suggest (ADDRESS_SPACE_END + 1 - SZ_1G) to help avoiding this >>> impression? >> >> I will jump on this to say that I am finding quite difficult to >> review >> code using define/variable that contains "end" in their name because >> it >> is never clear whether this is inclusive or exclusive. >> >> In this case, it is inclusive, but within the same patch I can see >> map_end which is exclusive. >> >> For this case, my suggestion would be to remove ADDRESS_SPACE_END and >> just hardcode the value where you want to put Xen. For others, you >> could >> use (start, size) to describe a region. > Thanks for the suggestion. I'll take it into account. >> >> Also, are you sure that all the CPU will be able to support more full >> 64-bit VA space? > I am not sure but based on Linux it looks like it is true. ( at least, > they use the same definitions for RV64 ). Hmm, the spec has this text in a rationalizing remark: "The mapping between 64-bit virtual addresses and the 39-bit usable address space of Sv39 is not based on zero-extension but instead follows an entrenched convention that allows an OS to use one or a few of the most-significant bits of a full-size (64-bit) virtual address to quickly distinguish user and supervisor address regions." Apart from that all descriptions look to assume that the top VA bits are simply ignored for the purpose of translating to PA, which would mean that any address within the 64-bit address space can be mapped, and the same mapping would appear many times in VA space. I wonder whether that's really how it's meant to be, or whether I'm overlooking something, or whether the spec would want clarifying. > I am curious how that can be checked. Try it out somewhere? Jan
On Wed, 2023-03-29 at 14:20 +0200, Jan Beulich wrote: > On 29.03.2023 13:47, Oleksii wrote: > > On Tue, 2023-03-28 at 16:44 +0100, Julien Grall wrote: > > > On 28/03/2023 16:34, Jan Beulich wrote: > > > > On 27.03.2023 19:17, Oleksii Kurochko wrote: > > > > > --- a/xen/arch/riscv/include/asm/config.h > > > > > +++ b/xen/arch/riscv/include/asm/config.h > > > > > @@ -39,12 +39,25 @@ > > > > > name: > > > > > #endif > > > > > > > > > > -#define XEN_VIRT_START _AT(UL, 0x80200000) > > > > > +#define ADDRESS_SPACE_END (_AC(-1, UL)) > > > > > +#define SZ_1G 0x40000000 > > > > > > > > No need for this - please use GB(1) (see xen/config.h) in its > > > > stead. > > > > > > > > > +#ifdef CONFIG_RISCV_64 > > > > > +#define XEN_VIRT_START (ADDRESS_SPACE_END - SZ_1G + 1) > > > > > > > > I first thought the " + 1" would be rendering the address > > > > misaligned. > > > > May I suggest (ADDRESS_SPACE_END + 1 - SZ_1G) to help avoiding > > > > this > > > > impression? > > > > > > I will jump on this to say that I am finding quite difficult to > > > review > > > code using define/variable that contains "end" in their name > > > because > > > it > > > is never clear whether this is inclusive or exclusive. > > > > > > In this case, it is inclusive, but within the same patch I can > > > see > > > map_end which is exclusive. > > > > > > For this case, my suggestion would be to remove ADDRESS_SPACE_END > > > and > > > just hardcode the value where you want to put Xen. For others, > > > you > > > could > > > use (start, size) to describe a region. > > Thanks for the suggestion. I'll take it into account. > > > > > > Also, are you sure that all the CPU will be able to support more > > > full > > > 64-bit VA space? > > I am not sure but based on Linux it looks like it is true. ( at > > least, > > they use the same definitions for RV64 ). > > Hmm, the spec has this text in a rationalizing remark: "The mapping > between 64-bit virtual addresses and the 39-bit usable address space > of Sv39 is not based on zero-extension but instead follows an > entrenched convention that allows an OS to use one or a few of the > most-significant bits of a full-size (64-bit) virtual address to > quickly distinguish user and supervisor address regions." > > Apart from that all descriptions look to assume that the top VA bits > are simply ignored for the purpose of translating to PA, which would > mean that any address within the 64-bit address space can be mapped, > and the same mapping would appear many times in VA space. I wonder > whether that's really how it's meant to be, or whether I'm > overlooking > something, or whether the spec would want clarifying. > You are right the top VA bits are ignored but my understanding is that you can't map any address within 64-bit address space as the top bits are ignored. Thereby address 0xFFFFFFXXXXXXXXXX are equal to 0x000000XXXXXXXXXX from HW point of view and the top VA bits are used only by developer. > > I am curious how that can be checked. > > Try it out somewhere? I re-read the initial question and also your answer and it looks like I misunderstood a little bit the initial question. So if the question was about if all CPUs support all MMU addressing modes ( Sv32, Sv39, Sv48, Sv57 ) then an answer is no. There is a rule that if ,for example, CPU supports Sv57 then it must support lower modes. If I understood the question correctly then we can add function which will write/read SATP_MODE: csr_write(CSR_SATP, ((unsigned long)stage1_pgtbl_root >> PAGE_SHIFT) | RV_STAGE1_MODE << SATP_MODE_SHIFT); __sfence_vma_all(); csr_write(CSR_SATP, satp); if ((csr_read(CSR_SATP) >> SATP_MODE_SHIFT) == SATP_MODE_SV57) { max_supported_mode = SATP_MODE_SV57; return } ... check other modes ... ~ Oleksii
On Wed, 2023-03-29 at 14:06 +0200, Jan Beulich wrote: > On 29.03.2023 13:43, Oleksii wrote: > > On Tue, 2023-03-28 at 17:34 +0200, Jan Beulich wrote: > > > On 27.03.2023 19:17, Oleksii Kurochko wrote: > > > > --- /dev/null > > > > +++ b/xen/arch/riscv/mm.c > > > > @@ -0,0 +1,277 @@ > > > > +#include <xen/init.h> > > > > +#include <xen/kernel.h> > > > > + > > > > +#include <asm/early_printk.h> > > > > +#include <asm/config.h> > > > > +#include <asm/csr.h> > > > > +#include <asm/mm.h> > > > > +#include <asm/page.h> > > > > +#include <asm/processor.h> > > > > + > > > > +#define PGTBL_INITIAL_COUNT 8 > > > > > > What does this magic number derive from? (A comment may help.) > > It can be now 4 tables as it is enough to map 2Mb. 8 it was when I > > experimented with bigger sizes of Xen binary and verified that > > logic of > > working with page tables works. > > Since this is connected to Xen's image size as you say, I guess the > constant wants to move to a header, and then be used in an assertion > in xen.lds.S. That way one will become easily aware if/when this 2Mb > assumption breaks. then we can not be tied to 2MB but to: ASSERT(_end - _start <= MB((PAGE_TABLE_INIT_COUNT - 2) * PAGE_SIZE), "Xen too large for early-boot assumptions") > > > > > +struct mmu_desc { > > > > + unsigned long num_levels; > > > > + uint32_t pgtbl_count; > > > > > > Nit: Style (as before please avoid fixed width types when their > > > use > > > isn't really warranted; afaics unsigned int will do fine here and > > > elsewhere below). > > I will change it but I am not sure that I fully understand what is > > an > > issue with uint32_t. > > The issue is simply that ./CODING_STYLE says to prefer basic types > whenever possible. Thanks. > > > > > +void __init setup_initial_pagetables(void) > > > > +{ > > > > + struct mmu_desc mmu_desc = { 0, 0, NULL, 0 }; > > > > + > > > > + /* > > > > + * Access to _{stard, end } is always PC-relative > > > > + * thereby when access them we will get load adresses > > > > + * of start and end of Xen > > > > + * To get linker addresses LOAD_TO_LINK() is required > > > > + * to use > > > > + */ > > > > + unsigned long load_start = (unsigned long)_start; > > > > + unsigned long load_end = (unsigned long)_end; > > > > + unsigned long linker_start = LOAD_TO_LINK(load_start); > > > > + unsigned long linker_end = LOAD_TO_LINK(load_end); > > > > + > > > > + if ( (linker_start != load_start) && > > > > + (linker_start <= load_end) && (load_start <= > > > > linker_end) > > > > ) { > > > > + early_printk("(XEN) linker and load address ranges > > > > overlap\n"); > > > > + die(); > > > > + } > > > > + > > > > + calc_pgtbl_lvls_num(&mmu_desc); > > > > + > > > > + mmu_desc.pgtbl_base = stage1_pgtbl_root; > > > > + mmu_desc.next_pgtbl = stage1_pgtbl_nonroot; > > > > + > > > > + setup_initial_mapping(&mmu_desc, > > > > + linker_start, > > > > + linker_end, > > > > + load_start, > > > > + PTE_LEAF_DEFAULT); > > > > + > > > > + setup_ptes_permission(&mmu_desc); > > > > > > ...: Why does this require a 2nd pass / function in the first > > > place? > > Probably I misunderstood Julien and it setup_pte_permission can be > > done > > in setup_initial_mapping() but here is the reply: > > https://lore.kernel.org/xen-devel/79e83610-5980-d9b5-7994-6b0cb2b9049a@xen.org/ > > Hmm, yes, his option 2 looks like what you've implemented. Still I > don't see why the permissions can't be got right on the first pass. ~ Oleksii
Hi Julien, On Wed, 2023-03-29 at 14:06 +0200, Jan Beulich wrote: > > > > +void __init setup_initial_pagetables(void) > > > > +{ > > > > + struct mmu_desc mmu_desc = { 0, 0, NULL, 0 }; > > > > + > > > > + /* > > > > + * Access to _{stard, end } is always PC-relative > > > > + * thereby when access them we will get load adresses > > > > + * of start and end of Xen > > > > + * To get linker addresses LOAD_TO_LINK() is required > > > > + * to use > > > > + */ > > > > + unsigned long load_start = (unsigned long)_start; > > > > + unsigned long load_end = (unsigned long)_end; > > > > + unsigned long linker_start = LOAD_TO_LINK(load_start); > > > > + unsigned long linker_end = LOAD_TO_LINK(load_end); > > > > + > > > > + if ( (linker_start != load_start) && > > > > + (linker_start <= load_end) && (load_start <= > > > > linker_end) > > > > ) { > > > > + early_printk("(XEN) linker and load address ranges > > > > overlap\n"); > > > > + die(); > > > > + } > > > > + > > > > + calc_pgtbl_lvls_num(&mmu_desc); > > > > + > > > > + mmu_desc.pgtbl_base = stage1_pgtbl_root; > > > > + mmu_desc.next_pgtbl = stage1_pgtbl_nonroot; > > > > + > > > > + setup_initial_mapping(&mmu_desc, > > > > + linker_start, > > > > + linker_end, > > > > + load_start, > > > > + PTE_LEAF_DEFAULT); > > > > + > > > > + setup_ptes_permission(&mmu_desc); > > > > > > ...: Why does this require a 2nd pass / function in the first > > > place? > > Probably I misunderstood Julien and it setup_pte_permission can be > > done > > in setup_initial_mapping() but here is the reply: > > https://lore.kernel.org/xen-devel/79e83610-5980-d9b5-7994-6b0cb2b9049a@xen.org/ > > Hmm, yes, his option 2 looks like what you've implemented. Still I > don't see why the permissions can't be got right on the first pass. I would like to ask you again about separation of mapping Xen and setting permission for specific sections. I am not using setup_initial_mapping to change permission flags as it was before so do we still need to do two passes? Can't we set permission during the first pass? It looks like there is a sense in merging setup_initial_mapping() and setup_ptes_permission(). Probably I misunderstood your opinion from the link above. ~ Oleksii
diff --git a/xen/arch/riscv/Makefile b/xen/arch/riscv/Makefile index 443f6bf15f..956ceb02df 100644 --- a/xen/arch/riscv/Makefile +++ b/xen/arch/riscv/Makefile @@ -1,5 +1,6 @@ obj-$(CONFIG_EARLY_PRINTK) += early_printk.o obj-y += entry.o +obj-y += mm.o obj-$(CONFIG_RISCV_64) += riscv64/ obj-y += sbi.o obj-y += setup.o diff --git a/xen/arch/riscv/include/asm/config.h b/xen/arch/riscv/include/asm/config.h index 763a922a04..7ffe9051b2 100644 --- a/xen/arch/riscv/include/asm/config.h +++ b/xen/arch/riscv/include/asm/config.h @@ -39,12 +39,25 @@ name: #endif -#define XEN_VIRT_START _AT(UL, 0x80200000) +#define ADDRESS_SPACE_END (_AC(-1, UL)) +#define SZ_1G 0x40000000 + +#ifdef CONFIG_RISCV_64 +#define XEN_VIRT_START (ADDRESS_SPACE_END - SZ_1G + 1) +#else +#error "RV32 isn't supported"asm +#endif #define SMP_CACHE_BYTES (1 << 6) #define STACK_SIZE PAGE_SIZE +#ifdef CONFIG_RISCV_64 +#define RV_STAGE1_MODE SATP_MODE_SV39 +#else +#define RV_STAGE1_MODE SATP_MODE_SV32 +#endif + #endif /* __RISCV_CONFIG_H__ */ /* * Local variables: diff --git a/xen/arch/riscv/include/asm/mm.h b/xen/arch/riscv/include/asm/mm.h new file mode 100644 index 0000000000..e16ce66fae --- /dev/null +++ b/xen/arch/riscv/include/asm/mm.h @@ -0,0 +1,9 @@ +#ifndef _ASM_RISCV_MM_H +#define _ASM_RISCV_MM_H + +void setup_initial_pagetables(void); + +void enable_mmu(void); +void cont_after_mmu_is_enabled(void); + +#endif /* _ASM_RISCV_MM_H */ diff --git a/xen/arch/riscv/include/asm/page-bits.h b/xen/arch/riscv/include/asm/page-bits.h index 1801820294..0879a527f2 100644 --- a/xen/arch/riscv/include/asm/page-bits.h +++ b/xen/arch/riscv/include/asm/page-bits.h @@ -1,6 +1,16 @@ #ifndef __RISCV_PAGE_BITS_H__ #define __RISCV_PAGE_BITS_H__ +#ifdef CONFIG_RISCV_64 +#define PAGETABLE_ORDER (9) +#else /* CONFIG_RISCV_32 */ +#define PAGETABLE_ORDER (10) +#endif + +#define PAGETABLE_ENTRIES (1 << PAGETABLE_ORDER) + +#define PTE_PPN_SHIFT 10 + #define PAGE_SHIFT 12 /* 4 KiB Pages */ #define PADDR_BITS 56 /* 44-bit PPN */ diff --git a/xen/arch/riscv/include/asm/page.h b/xen/arch/riscv/include/asm/page.h new file mode 100644 index 0000000000..f1dd5142cf --- /dev/null +++ b/xen/arch/riscv/include/asm/page.h @@ -0,0 +1,62 @@ +#ifndef _ASM_RISCV_PAGE_H +#define _ASM_RISCV_PAGE_H + +#include <xen/const.h> +#include <xen/types.h> + +#define VPN_MASK ((unsigned long)(PAGETABLE_ENTRIES - 1)) + +#define XEN_PT_LEVEL_ORDER(lvl) ((lvl) * PAGETABLE_ORDER) +#define XEN_PT_LEVEL_SHIFT(lvl) (XEN_PT_LEVEL_ORDER(lvl) + PAGE_SHIFT) +#define XEN_PT_LEVEL_SIZE(lvl) (_AT(paddr_t, 1) << XEN_PT_LEVEL_SHIFT(lvl)) +#define XEN_PT_LEVEL_MAP_MASK(lvl) (~(XEN_PT_LEVEL_SIZE(lvl) - 1)) +#define XEN_PT_LEVEL_MASK(lvl) (VPN_MASK << XEN_PT_LEVEL_SHIFT(lvl)) + +#define PTE_VALID BIT(0, UL) +#define PTE_READABLE BIT(1, UL) +#define PTE_WRITABLE BIT(2, UL) +#define PTE_EXECUTABLE BIT(3, UL) +#define PTE_USER BIT(4, UL) +#define PTE_GLOBAL BIT(5, UL) +#define PTE_ACCESSED BIT(6, UL) +#define PTE_DIRTY BIT(7, UL) +#define PTE_RSW (BIT(8, UL) | BIT(9, UL)) + +#define PTE_LEAF_DEFAULT (PTE_VALID | PTE_READABLE | PTE_WRITABLE) +#define PTE_TABLE (PTE_VALID) + +/* Calculate the offsets into the pagetables for a given VA */ +#define pt_linear_offset(lvl, va) ((va) >> XEN_PT_LEVEL_SHIFT(lvl)) + +#define pt_index(lvl, va) pt_linear_offset(lvl, (va) & XEN_PT_LEVEL_MASK(lvl)) + +#ifdef CONFIG_RISCV_64 +typedef uint64_t pte_t; +#else +typedef uint32_t pte_t; +#endif + +#define addr_to_pte(x) (((x) >> PTE_PPN_SHIFT) << PAGE_SHIFT) + +/* Shift the VPN[x] or PPN[x] fields of a virtual or physical address + * to become the shifted PPN[x] fields of a page table entry */ +#define addr_to_ppn(x) (((x) >> PAGE_SHIFT) << PTE_PPN_SHIFT) + +static inline pte_t paddr_to_pte(const unsigned long paddr, + const unsigned long permissions) +{ + pte_t tmp = (pte_t)addr_to_ppn(paddr); + return tmp | permissions; +} + +static inline pte_t pte_to_paddr(const unsigned long paddr) +{ + return (pte_t)addr_to_pte(paddr); +} + +static inline bool pte_is_valid(const pte_t p) +{ + return p & PTE_VALID; +} + +#endif /* _ASM_RISCV_PAGE_H */ diff --git a/xen/arch/riscv/mm.c b/xen/arch/riscv/mm.c new file mode 100644 index 0000000000..2179ae9fd4 --- /dev/null +++ b/xen/arch/riscv/mm.c @@ -0,0 +1,277 @@ +#include <xen/init.h> +#include <xen/kernel.h> + +#include <asm/early_printk.h> +#include <asm/config.h> +#include <asm/csr.h> +#include <asm/mm.h> +#include <asm/page.h> +#include <asm/processor.h> + +#define PGTBL_INITIAL_COUNT 8 +#define PGTBL_ENTRY_AMOUNT (PAGE_SIZE/sizeof(pte_t)) + +struct mmu_desc { + unsigned long num_levels; + uint32_t pgtbl_count; + pte_t *next_pgtbl; + pte_t *pgtbl_base; +}; + +extern unsigned long __bss_start[]; +extern unsigned long __bss_end[]; +extern unsigned char cpu0_boot_stack[STACK_SIZE]; + +#define PHYS_OFFSET ((unsigned long)_start - XEN_VIRT_START) +#define LOAD_TO_LINK(addr) ((addr) - PHYS_OFFSET) +#define LINK_TO_LOAD(addr) ((addr) + PHYS_OFFSET) + +pte_t __section(".bss.page_aligned") __aligned(PAGE_SIZE) +stage1_pgtbl_root[PGTBL_ENTRY_AMOUNT]; + +pte_t __section(".bss.page_aligned") __aligned(PAGE_SIZE) +stage1_pgtbl_nonroot[PGTBL_INITIAL_COUNT * PGTBL_ENTRY_AMOUNT]; + +#define HANDLE_PGTBL() \ + index = pt_index(curr_lvl_num, page_addr); \ + if ( pte_is_valid(pgtbl[index]) ) { \ + /* Find L{ 0-3 } table */ \ + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); \ + } else { \ + /* Allocate new L{0-3} page table */ \ + if ( mmu_desc->pgtbl_count == PGTBL_INITIAL_COUNT ) { \ + early_printk("(XEN) No initial table available\n"); \ + /* panic(), BUG() or ASSERT() aren't ready now. */ \ + die(); \ + } \ + mmu_desc->pgtbl_count++; \ + pgtbl[index] = paddr_to_pte((unsigned long)mmu_desc->next_pgtbl, \ + PTE_VALID); \ + pgtbl = mmu_desc->next_pgtbl; \ + mmu_desc->next_pgtbl += PGTBL_ENTRY_AMOUNT; \ + } + +static void __init setup_initial_mapping(struct mmu_desc *mmu_desc, + unsigned long map_start, + unsigned long map_end, + unsigned long pa_start, + unsigned long permission) +{ + uint32_t index; + pte_t *pgtbl; + unsigned long page_addr; + unsigned int curr_lvl_num; + pte_t pte_to_be_written; + unsigned long paddr; + + if ( (unsigned long)_start % XEN_PT_LEVEL_SIZE(0) ) { + early_printk("(XEN) Xen should be loaded at 4k boundary\n"); + die(); + } + + if ( map_start & ~XEN_PT_LEVEL_MAP_MASK(0) || + pa_start & ~XEN_PT_LEVEL_MAP_MASK(0) ) { + + early_printk("(XEN) map and pa start addresses should be aligned\n"); + /* panic(), BUG() or ASSERT() aren't ready now. */ + die(); + } + + page_addr = map_start; + while ( page_addr < map_end ) { + pgtbl = mmu_desc->pgtbl_base; + + switch (mmu_desc->num_levels) + { + case 4: /* Level 3 */ + curr_lvl_num = 3; + HANDLE_PGTBL(); + case 3: /* Level 2 */ + curr_lvl_num = 2; + HANDLE_PGTBL(); + case 2: /* Level 1 */ + curr_lvl_num = 1; + HANDLE_PGTBL(); + case 1: /* Level 0 */ + index = pt_index(0, page_addr); + paddr = (page_addr - map_start) + pa_start; + pte_to_be_written = paddr_to_pte(paddr, permission); + + if ( !pte_is_valid(pgtbl[index]) ) + pgtbl[index] = pte_to_be_written; + else { + /* + * get an adresses of current pte and that one to + * be written without permission flags + */ + unsigned long curr_pte = + (unsigned long)&pgtbl[index] & (PTE_PPN_SHIFT - 1); + pte_to_be_written &= (PTE_PPN_SHIFT - 1); + + if ( curr_pte != pte_to_be_written ) { + early_printk("PTE that is intended to write isn't the \ + same that the once are overwriting\n"); + /* panic(), <asm/bug.h> aren't ready now. */ + die(); + } + } + } + + /* Point to next page */ + page_addr += XEN_PT_LEVEL_SIZE(0); + } +} + +static void __init calc_pgtbl_lvls_num(struct mmu_desc *mmu_desc) +{ + unsigned long satp_mode = RV_STAGE1_MODE; + + /* Number of page table levels */ + switch (satp_mode) { + case SATP_MODE_SV32: + mmu_desc->num_levels = 2; + break; + case SATP_MODE_SV39: + mmu_desc->num_levels = 3; + break; + case SATP_MODE_SV48: + mmu_desc->num_levels = 4; + break; + default: + early_printk("(XEN) Unsupported SATP_MODE\n"); + while (1); + } +} + +static void __init setup_ptes_permission(struct mmu_desc *mmu_desc) +{ + pte_t *pgtbl; + uint32_t index; + uint32_t permissions = PTE_VALID; + unsigned long start = LOAD_TO_LINK((unsigned long)_start); + unsigned long end = LOAD_TO_LINK((unsigned long)_end); + + while ( start < end ) + { + pgtbl = (pte_t *) mmu_desc->pgtbl_base; + + switch (mmu_desc->num_levels) + { + case 4: /* find Level 3 page table*/ + index = pt_index(3, start); + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); + case 3: /* find level 2 page table */ + index = pt_index(2, start); + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); + case 2: /* find level 1 page table */ + index = pt_index(1, start); + pgtbl = (pte_t *)pte_to_paddr(pgtbl[index]); + case 1: /* find level 0 page table */ + index = pt_index(0, start); + + if ( is_kernel_text(LINK_TO_LOAD(start)) || + is_kernel_inittext(LINK_TO_LOAD(start)) ) + permissions |= PTE_EXECUTABLE | PTE_READABLE; + + if ( is_kernel_rodata(LINK_TO_LOAD(start)) ) + permissions |= PTE_READABLE; + + pgtbl[index] |= permissions; + } + + start += PAGE_SIZE; + } +} + +/* + * setup_initial_pagetables: + * + * Build the page tables for Xen that map the following: + * 1. Calculate page table's level numbers. + * 2. Init mmu description structure. + * 3. Check that linker addresses range doesn't overlap + * with load addresses range + * 4. Map all linker addresses and load addresses ( it shouldn't + * be 1:1 mapped and will be 1:1 mapped only in case if + * linker address is equal to load address ) with + * RW permissions by default. + * 5. Setup proper PTE permissions for each section. + */ +void __init setup_initial_pagetables(void) +{ + struct mmu_desc mmu_desc = { 0, 0, NULL, 0 }; + + /* + * Access to _{stard, end } is always PC-relative + * thereby when access them we will get load adresses + * of start and end of Xen + * To get linker addresses LOAD_TO_LINK() is required + * to use + */ + unsigned long load_start = (unsigned long)_start; + unsigned long load_end = (unsigned long)_end; + unsigned long linker_start = LOAD_TO_LINK(load_start); + unsigned long linker_end = LOAD_TO_LINK(load_end); + + if ( (linker_start != load_start) && + (linker_start <= load_end) && (load_start <= linker_end) ) { + early_printk("(XEN) linker and load address ranges overlap\n"); + die(); + } + + calc_pgtbl_lvls_num(&mmu_desc); + + mmu_desc.pgtbl_base = stage1_pgtbl_root; + mmu_desc.next_pgtbl = stage1_pgtbl_nonroot; + + setup_initial_mapping(&mmu_desc, + linker_start, + linker_end, + load_start, + PTE_LEAF_DEFAULT); + + setup_ptes_permission(&mmu_desc); +} + +void __init enable_mmu() +{ + /* + * Calculate a linker time address of the mmu_is_enabled + * label and update CSR_STVEC with it. + * MMU is configured in a way where linker addresses are mapped + * on load addresses so in a case when linker addresses are not equal + * to load addresses, after MMU is enabled, it will cause + * an exception and jump to linker time addresses. + * Otherwise if load addresses are equal to linker addresses the code + * after mmu_is_enabled label will be executed without exception. + */ + csr_write(CSR_STVEC, LOAD_TO_LINK((unsigned long)&&mmu_is_enabled)); + + /* Ensure page table writes precede loading the SATP */ + asm volatile("sfence.vma"); + + /* Enable the MMU and load the new pagetable for Xen */ + csr_write(CSR_SATP, + ((unsigned long)stage1_pgtbl_root >> PAGE_SHIFT) | RV_STAGE1_MODE << SATP_MODE_SHIFT); + + asm volatile(".align 2"); +mmu_is_enabled: + /* + * Stack should be re-inited as: + * 1. Right now an address of the stack is relative to load time + * addresses what will cause an issue in case of load start address + * isn't equal to linker start address. + * 2. Addresses in stack are all load time relative which can be an + * issue in case when load start address isn't equal to linker + * start address. + */ + asm volatile ("mv sp, %0" : : "r"((unsigned long)cpu0_boot_stack + STACK_SIZE)); + + /* + * We can't return to the caller because the stack was reseted + * and it may have stash some variable on the stack. + * Jump to a brand new function as the stack was reseted + */ + cont_after_mmu_is_enabled(); +} + diff --git a/xen/arch/riscv/riscv64/head.S b/xen/arch/riscv/riscv64/head.S index 8887f0cbd4..b3309d902c 100644 --- a/xen/arch/riscv/riscv64/head.S +++ b/xen/arch/riscv/riscv64/head.S @@ -1,4 +1,5 @@ #include <asm/asm.h> +#include <asm/asm-offsets.h> #include <asm/riscv_encoding.h> .section .text.header, "ax", %progbits @@ -32,3 +33,4 @@ ENTRY(start) add sp, sp, t0 tail start_xen + diff --git a/xen/arch/riscv/setup.c b/xen/arch/riscv/setup.c index 3786f337e0..315804aa87 100644 --- a/xen/arch/riscv/setup.c +++ b/xen/arch/riscv/setup.c @@ -2,6 +2,7 @@ #include <xen/init.h> #include <asm/early_printk.h> +#include <asm/mm.h> /* Xen stack for bringing up the first CPU. */ unsigned char __initdata cpu0_boot_stack[STACK_SIZE] @@ -26,3 +27,13 @@ void __init noreturn start_xen(unsigned long bootcpu_id, unreachable(); } + +void __init noreturn cont_after_mmu_is_enabled(void) +{ + early_printk("All set up\n"); + + for ( ;; ) + asm volatile ("wfi"); + + unreachable(); +} diff --git a/xen/arch/riscv/xen.lds.S b/xen/arch/riscv/xen.lds.S index f299ea8422..54d4236866 100644 --- a/xen/arch/riscv/xen.lds.S +++ b/xen/arch/riscv/xen.lds.S @@ -136,6 +136,7 @@ SECTIONS . = ALIGN(POINTER_ALIGN); __init_end = .; + . = ALIGN(PAGE_SIZE); .bss : { /* BSS */ __bss_start = .; *(.bss.stack_aligned) @@ -169,3 +170,5 @@ SECTIONS ASSERT(!SIZEOF(.got), ".got non-empty") ASSERT(!SIZEOF(.got.plt), ".got.plt non-empty") + +ASSERT(_end - _start <= MB(2), "Xen too large for early-boot assumptions")
The idea was taken from xvisor but the following changes were done: * Use only a minimal part of the code enough to enable MMU * rename {_}setup_initial_pagetables functions * add an argument for setup_initial_mapping to have an opportunity to make set PTE flags. * update setup_initial_pagetables function to map sections with correct PTE flags. * Rewrite enable_mmu() to C. * map linker addresses range to load addresses range without 1:1 mapping. It will be 1:1 only in case when load_start_addr is equal to linker_start_addr. * add safety checks such as: * Xen size is less than page size * linker addresses range doesn't overlap load addresses range * Rework macros {THIRD,SECOND,FIRST,ZEROETH}_{SHIFT,MASK} * change PTE_LEAF_DEFAULT to RW instead of RWX. * Remove phys_offset as it is not used now * Remove alignment of {map, pa}_start &= XEN_PT_LEVEL_MAP_MASK(0); in setup_inital_mapping() as they should be already aligned. Make a check that {map_pa}_start are aligned. * Remove clear_pagetables() as initial pagetables will be zeroed during bss initialization * Remove __attribute__((section(".entry")) for setup_initial_pagetables() as there is no such section in xen.lds.S * Update the argument of pte_is_valid() to "const pte_t *p" * Add check that Xen's load address is aligned at 4k boundary * Refactor setup_initial_pagetables() so it is mapping linker address range to load address range. After setup needed permissions for specific section ( such as .text, .rodata, etc ) otherwise RW permission will be set by default. Origin: git@github.com:xvisor/xvisor.git 9be2fdd7 Signed-off-by: Oleksii Kurochko <oleksii.kurochko@gmail.com> --- Changes in V3: - update definition of pte_t structure to have a proper size of pte_t in case of RV32. - update asm/mm.h with new functions and remove unnecessary 'extern'. - remove LEVEL_* macros as only XEN_PT_LEVEL_* are enough. - update paddr_to_pte() to receive permissions as an argument. - add check that map_start & pa_start is properly aligned. - move defines PAGETABLE_ORDER, PAGETABLE_ENTRIES, PTE_PPN_SHIFT to <asm/page-bits.h> - Rename PTE_SHIFT to PTE_PPN_SHIFT - refactor setup_initial_pagetables: map all LINK addresses to LOAD addresses and after setup PTEs permission for sections; update check that linker and load addresses don't overlap. - refactor setup_initial_mapping: allocate pagetable 'dynamically' if it is necessary. - rewrite enable_mmu in C; add the check that map_start and pa_start are aligned on 4k boundary. - update the comment for setup_initial_pagetable funcion - Add RV_STAGE1_MODE to support different MMU modes - set XEN_VIRT_START very high to not overlap with load address range - align bss section --- Changes in V2: * update the commit message: * Remove {ZEROETH,FIRST,...}_{SHIFT,MASK, SIZE,...} and introduce instead of them XEN_PT_LEVEL_*() and LEVEL_* * Rework pt_linear_offset() and pt_index based on XEN_PT_LEVEL_*() * Remove clear_pagetables() functions as pagetables were zeroed during .bss initialization * Rename _setup_initial_pagetables() to setup_initial_mapping() * Make PTE_DEFAULT equal to RX. * Update prototype of setup_initial_mapping(..., bool writable) -> setup_initial_mapping(..., UL flags) * Update calls of setup_initial_mapping according to new prototype * Remove unnecessary call of: _setup_initial_pagetables(..., load_addr_start, load_addr_end, load_addr_start, ...) * Define index* in the loop of setup_initial_mapping * Remove attribute "__attribute__((section(".entry")))" for setup_initial_pagetables() as we don't have such section * make arguments of paddr_to_pte() and pte_is_valid() as const. * make xen_second_pagetable static. * use <xen/kernel.h> instead of declaring extern unsigned long _stext, 0etext, _srodata, _erodata * update 'extern unsigned long __init_begin' to 'extern unsigned long __init_begin[]' * use aligned() instead of "__attribute__((__aligned__(PAGE_SIZE)))" * set __section(".bss.page_aligned") for page tables arrays * fix identatations * Change '__attribute__((section(".entry")))' to '__init' * Remove phys_offset as it isn't used now. * Remove alignment of {map, pa}_start &= XEN_PT_LEVEL_MAP_MASK(0); in setup_inital_mapping() as they should be already aligned. * Remove clear_pagetables() as initial pagetables will be zeroed during bss initialization * Remove __attribute__((section(".entry")) for setup_initial_pagetables() as there is no such section in xen.lds.S * Update the argument of pte_is_valid() to "const pte_t *p" --- xen/arch/riscv/Makefile | 1 + xen/arch/riscv/include/asm/config.h | 15 +- xen/arch/riscv/include/asm/mm.h | 9 + xen/arch/riscv/include/asm/page-bits.h | 10 + xen/arch/riscv/include/asm/page.h | 62 ++++++ xen/arch/riscv/mm.c | 277 +++++++++++++++++++++++++ xen/arch/riscv/riscv64/head.S | 2 + xen/arch/riscv/setup.c | 11 + xen/arch/riscv/xen.lds.S | 3 + 9 files changed, 389 insertions(+), 1 deletion(-) create mode 100644 xen/arch/riscv/include/asm/mm.h create mode 100644 xen/arch/riscv/include/asm/page.h create mode 100644 xen/arch/riscv/mm.c