@@ -79,7 +79,7 @@ struct hash_pte {
typedef struct {
unsigned long id;
- unsigned long vdso_base;
+ unsigned long vdso;
} mm_context_t;
#endif /* !__ASSEMBLY__ */
@@ -525,7 +525,7 @@ typedef struct {
#else
u16 sllp; /* SLB page size encoding */
#endif
- unsigned long vdso_base;
+ unsigned long vdso;
#ifdef CONFIG_PPC_SUBPAGE_PROT
struct subpage_prot_table spt;
#endif /* CONFIG_PPC_SUBPAGE_PROT */
@@ -18,10 +18,10 @@ static inline void arch_remap(struct mm_struct *mm,
{
/*
* mremap() doesn't allow moving multiple vmas so we can limit the
- * check to old_start == vdso_base.
+ * check to old_start == vdso.
*/
- if (old_start == mm->context.vdso_base)
- mm->context.vdso_base = new_start;
+ if (old_start == mm->context.vdso)
+ mm->context.vdso = new_start;
}
#define arch_remap arch_remap
@@ -56,7 +56,7 @@
typedef struct {
unsigned int id;
unsigned int active;
- unsigned long vdso_base;
+ unsigned long vdso;
} mm_context_t;
#endif /* !__ASSEMBLY__ */
@@ -107,7 +107,7 @@ extern unsigned int tlb_44x_index;
typedef struct {
unsigned int id;
unsigned int active;
- unsigned long vdso_base;
+ unsigned long vdso;
} mm_context_t;
#endif /* !__ASSEMBLY__ */
@@ -167,7 +167,7 @@
typedef struct {
unsigned int id;
unsigned int active;
- unsigned long vdso_base;
+ unsigned long vdso;
} mm_context_t;
#endif /* !__ASSEMBLY__ */
@@ -228,7 +228,7 @@ extern unsigned int tlbcam_index;
typedef struct {
unsigned int id;
unsigned int active;
- unsigned long vdso_base;
+ unsigned long vdso;
#ifdef CONFIG_PPC_MM_SLICES
u64 low_slices_psize; /* SLB page size encodings */
u64 high_slices_psize; /* 4 bits per slice for now */
@@ -139,8 +139,8 @@ static inline void arch_unmap(struct mm_struct *mm,
struct vm_area_struct *vma,
unsigned long start, unsigned long end)
{
- if (start <= mm->context.vdso_base && mm->context.vdso_base < end)
- mm->context.vdso_base = 0;
+ if (start <= mm->context.vdso && mm->context.vdso < end)
+ mm->context.vdso = 0;
}
static inline void arch_bprm_mm_init(struct mm_struct *mm,
@@ -17,7 +17,7 @@
#ifndef __ASSEMBLY__
-/* Offsets relative to thread->vdso_base */
+/* Offsets relative to mm->context.vdso */
extern unsigned long vdso64_rt_sigtramp;
extern unsigned long vdso32_sigtramp;
extern unsigned long vdso32_rt_sigtramp;
@@ -177,7 +177,7 @@ do { \
NEW_AUX_ENT(AT_DCACHEBSIZE, dcache_bsize); \
NEW_AUX_ENT(AT_ICACHEBSIZE, icache_bsize); \
NEW_AUX_ENT(AT_UCACHEBSIZE, ucache_bsize); \
- VDSO_AUX_ENT(AT_SYSINFO_EHDR, current->mm->context.vdso_base); \
+ VDSO_AUX_ENT(AT_SYSINFO_EHDR, current->mm->context.vdso); \
} while (0)
/* PowerPC64 relocations defined by the ABIs */
@@ -1000,9 +1000,9 @@ int handle_rt_signal32(struct ksignal *ksig, sigset_t *oldset,
/* Save user registers on the stack */
frame = &rt_sf->uc.uc_mcontext;
addr = frame;
- if (vdso32_rt_sigtramp && current->mm->context.vdso_base) {
+ if (vdso32_rt_sigtramp && current->mm->context.vdso) {
sigret = 0;
- tramp = current->mm->context.vdso_base + vdso32_rt_sigtramp;
+ tramp = current->mm->context.vdso + vdso32_rt_sigtramp;
} else {
sigret = __NR_rt_sigreturn;
tramp = (unsigned long) frame->tramp;
@@ -1425,9 +1425,9 @@ int handle_signal32(struct ksignal *ksig, sigset_t *oldset, struct pt_regs *regs
|| __put_user(ksig->sig, &sc->signal))
goto badframe;
- if (vdso32_sigtramp && current->mm->context.vdso_base) {
+ if (vdso32_sigtramp && current->mm->context.vdso) {
sigret = 0;
- tramp = current->mm->context.vdso_base + vdso32_sigtramp;
+ tramp = current->mm->context.vdso + vdso32_sigtramp;
} else {
sigret = __NR_sigreturn;
tramp = (unsigned long) frame->mctx.tramp;
@@ -751,8 +751,8 @@ int handle_rt_signal64(struct ksignal *ksig, sigset_t *set, struct pt_regs *regs
current->thread.fp_state.fpscr = 0;
/* Set up to return from userspace. */
- if (vdso64_rt_sigtramp && current->mm->context.vdso_base) {
- regs->link = current->mm->context.vdso_base + vdso64_rt_sigtramp;
+ if (vdso64_rt_sigtramp && current->mm->context.vdso) {
+ regs->link = current->mm->context.vdso + vdso64_rt_sigtramp;
} else {
err |= setup_trampoline(__NR_rt_sigreturn, &frame->tramp[0]);
if (err)
@@ -179,7 +179,7 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp)
vdso_base = VDSO32_MBASE;
#endif
- current->mm->context.vdso_base = 0;
+ current->mm->context.vdso = 0;
/* vDSO has a problem and was disabled, just don't "enable" it for the
* process
@@ -213,7 +213,7 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp)
* install_special_mapping or the perf counter mmap tracking code
* will fail to recognise it as a vDSO (since arch_vma_name fails).
*/
- current->mm->context.vdso_base = vdso_base;
+ current->mm->context.vdso = vdso_base;
/*
* our vma flags don't have VM_WRITE so by default, the process isn't
@@ -230,7 +230,7 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp)
VM_MAYREAD|VM_MAYWRITE|VM_MAYEXEC,
vdso_pagelist);
if (rc) {
- current->mm->context.vdso_base = 0;
+ current->mm->context.vdso = 0;
goto fail_mmapsem;
}
@@ -244,7 +244,7 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp)
const char *arch_vma_name(struct vm_area_struct *vma)
{
- if (vma->vm_mm && vma->vm_start == vma->vm_mm->context.vdso_base)
+ if (vma->vm_mm && vma->vm_start == vma->vm_mm->context.vdso)
return "[vdso]";
return NULL;
}
@@ -209,8 +209,8 @@ static int is_sigreturn_64_address(unsigned long nip, unsigned long fp)
{
if (nip == fp + offsetof(struct signal_frame_64, tramp))
return 1;
- if (vdso64_rt_sigtramp && current->mm->context.vdso_base &&
- nip == current->mm->context.vdso_base + vdso64_rt_sigtramp)
+ if (vdso64_rt_sigtramp && current->mm->context.vdso &&
+ nip == current->mm->context.vdso + vdso64_rt_sigtramp)
return 1;
return 0;
}
@@ -368,8 +368,8 @@ static int is_sigreturn_32_address(unsigned int nip, unsigned int fp)
{
if (nip == fp + offsetof(struct signal_frame_32, mctx.mc_pad))
return 1;
- if (vdso32_sigtramp && current->mm->context.vdso_base &&
- nip == current->mm->context.vdso_base + vdso32_sigtramp)
+ if (vdso32_sigtramp && current->mm->context.vdso &&
+ nip == current->mm->context.vdso + vdso32_sigtramp)
return 1;
return 0;
}
@@ -379,8 +379,8 @@ static int is_rt_sigreturn_32_address(unsigned int nip, unsigned int fp)
if (nip == fp + offsetof(struct rt_signal_frame_32,
uc.uc_mcontext.mc_pad))
return 1;
- if (vdso32_rt_sigtramp && current->mm->context.vdso_base &&
- nip == current->mm->context.vdso_base + vdso32_rt_sigtramp)
+ if (vdso32_rt_sigtramp && current->mm->context.vdso &&
+ nip == current->mm->context.vdso + vdso32_rt_sigtramp)
return 1;
return 0;
}
In order to share remap and unmap support for the VDSO with other architectures without duplicating the code, we need a common name and type for the address of the VDSO. An informal survey of the architectures indicates unsigned long vdso is popular. Change the variable name in powerpc from mm->context.vdso_base to simply mm->context.vdso. Signed-off-by: Christopher Covington <cov@codeaurora.org> --- arch/powerpc/include/asm/book3s/32/mmu-hash.h | 2 +- arch/powerpc/include/asm/book3s/64/mmu-hash.h | 2 +- arch/powerpc/include/asm/mm-arch-hooks.h | 6 +++--- arch/powerpc/include/asm/mmu-40x.h | 2 +- arch/powerpc/include/asm/mmu-44x.h | 2 +- arch/powerpc/include/asm/mmu-8xx.h | 2 +- arch/powerpc/include/asm/mmu-book3e.h | 2 +- arch/powerpc/include/asm/mmu_context.h | 4 ++-- arch/powerpc/include/asm/vdso.h | 2 +- arch/powerpc/include/uapi/asm/elf.h | 2 +- arch/powerpc/kernel/signal_32.c | 8 ++++---- arch/powerpc/kernel/signal_64.c | 4 ++-- arch/powerpc/kernel/vdso.c | 8 ++++---- arch/powerpc/perf/callchain.c | 12 ++++++------ 14 files changed, 29 insertions(+), 29 deletions(-)