@@ -401,6 +401,7 @@ void destroy_context_ldt(struct mm_struct *mm)
free_ldt_struct(mm->context.ldt);
mm->context.ldt = NULL;
}
+EXPORT_SYMBOL_GPL(destroy_context_ldt);
void ldt_arch_exit_mmap(struct mm_struct *mm)
{
@@ -70,7 +70,7 @@ static void clear_asid_other(void)
}
atomic64_t last_mm_ctx_id = ATOMIC64_INIT(1);
-
+EXPORT_SYMBOL_GPL(last_mm_ctx_id);
static void choose_new_asid(struct mm_struct *next, u64 next_tlb_gen,
u16 *new_asid, bool *need_flush)
@@ -159,6 +159,7 @@ void switch_mm(struct mm_struct *prev, struct mm_struct *next,
switch_mm_irqs_off(prev, next, tsk);
local_irq_restore(flags);
}
+EXPORT_SYMBOL_GPL(switch_mm);
static void sync_current_stack_to_mm(struct mm_struct *mm)
{
@@ -434,6 +434,7 @@ int __pte_alloc(struct mm_struct *mm, pmd_t *pmd)
pte_free(mm, new);
return 0;
}
+EXPORT_SYMBOL_GPL(__pte_alloc);
int __pte_alloc_kernel(pmd_t *pmd)
{
@@ -453,6 +454,7 @@ int __pte_alloc_kernel(pmd_t *pmd)
pte_free_kernel(&init_mm, new);
return 0;
}
+EXPORT_SYMBOL_GPL(__pte_alloc_kernel);
static inline void init_rss_vec(int *rss)
{
@@ -4007,6 +4009,7 @@ int __p4d_alloc(struct mm_struct *mm, pgd_t *pgd, unsigned long address)
spin_unlock(&mm->page_table_lock);
return 0;
}
+EXPORT_SYMBOL_GPL(__p4d_alloc);
#endif /* __PAGETABLE_P4D_FOLDED */
#ifndef __PAGETABLE_PUD_FOLDED
@@ -4039,6 +4042,7 @@ int __pud_alloc(struct mm_struct *mm, p4d_t *p4d, unsigned long address)
spin_unlock(&mm->page_table_lock);
return 0;
}
+EXPORT_SYMBOL_GPL(__pud_alloc);
#endif /* __PAGETABLE_PUD_FOLDED */
#ifndef __PAGETABLE_PMD_FOLDED
@@ -4072,6 +4076,7 @@ int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address)
spin_unlock(ptl);
return 0;
}
+EXPORT_SYMBOL_GPL(__pmd_alloc);
#endif /* __PAGETABLE_PMD_FOLDED */
static int __follow_pte_pmd(struct mm_struct *mm, unsigned long address,