Message ID | 20210513184734.29317-2-rppt@kernel.org (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | mm: introduce memfd_secret system call to create "secret" memory areas | expand |
On 13.05.21 20:47, Mike Rapoport wrote: > From: Mike Rapoport <rppt@linux.ibm.com> > > It will be used by the upcoming secret memory implementation. > > Signed-off-by: Mike Rapoport <rppt@linux.ibm.com> > Cc: Alexander Viro <viro@zeniv.linux.org.uk> > Cc: Andy Lutomirski <luto@kernel.org> > Cc: Arnd Bergmann <arnd@arndb.de> > Cc: Borislav Petkov <bp@alien8.de> > Cc: Catalin Marinas <catalin.marinas@arm.com> > Cc: Christopher Lameter <cl@linux.com> > Cc: Dan Williams <dan.j.williams@intel.com> > Cc: Dave Hansen <dave.hansen@linux.intel.com> > Cc: David Hildenbrand <david@redhat.com> > Cc: Elena Reshetova <elena.reshetova@intel.com> > Cc: Hagen Paul Pfeifer <hagen@jauu.net> > Cc: "H. Peter Anvin" <hpa@zytor.com> > Cc: Ingo Molnar <mingo@redhat.com> > Cc: James Bottomley <jejb@linux.ibm.com> > Cc: "Kirill A. Shutemov" <kirill@shutemov.name> > Cc: Mark Rutland <mark.rutland@arm.com> > Cc: Matthew Wilcox <willy@infradead.org> > Cc: Michael Kerrisk <mtk.manpages@gmail.com> > Cc: Palmer Dabbelt <palmer@dabbelt.com> > Cc: Palmer Dabbelt <palmerdabbelt@google.com> > Cc: Paul Walmsley <paul.walmsley@sifive.com> > Cc: Peter Zijlstra <peterz@infradead.org> > Cc: Rick Edgecombe <rick.p.edgecombe@intel.com> > Cc: Roman Gushchin <guro@fb.com> > Cc: Shakeel Butt <shakeelb@google.com> > Cc: Shuah Khan <shuah@kernel.org> > Cc: Thomas Gleixner <tglx@linutronix.de> > Cc: Tycho Andersen <tycho@tycho.ws> > Cc: Will Deacon <will@kernel.org> > --- > mm/internal.h | 3 +++ > mm/mmap.c | 5 ++--- > 2 files changed, 5 insertions(+), 3 deletions(-) > > diff --git a/mm/internal.h b/mm/internal.h > index 54bd0dc2c23c..46eb82eaa195 100644 > --- a/mm/internal.h > +++ b/mm/internal.h > @@ -373,6 +373,9 @@ static inline void munlock_vma_pages_all(struct vm_area_struct *vma) > extern void mlock_vma_page(struct page *page); > extern unsigned int munlock_vma_page(struct page *page); > > +extern int mlock_future_check(struct mm_struct *mm, unsigned long flags, > + unsigned long len); > + > /* > * Clear the page's PageMlocked(). This can be useful in a situation where > * we want to unconditionally remove a page from the pagecache -- e.g., > diff --git a/mm/mmap.c b/mm/mmap.c > index 0584e540246e..81f5595a8490 100644 > --- a/mm/mmap.c > +++ b/mm/mmap.c > @@ -1352,9 +1352,8 @@ static inline unsigned long round_hint_to_min(unsigned long hint) > return hint; > } > > -static inline int mlock_future_check(struct mm_struct *mm, > - unsigned long flags, > - unsigned long len) > +int mlock_future_check(struct mm_struct *mm, unsigned long flags, > + unsigned long len) > { > unsigned long locked, lock_limit; > > Reviewed-by: David Hildenbrand <david@redhat.com>
diff --git a/mm/internal.h b/mm/internal.h index 54bd0dc2c23c..46eb82eaa195 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -373,6 +373,9 @@ static inline void munlock_vma_pages_all(struct vm_area_struct *vma) extern void mlock_vma_page(struct page *page); extern unsigned int munlock_vma_page(struct page *page); +extern int mlock_future_check(struct mm_struct *mm, unsigned long flags, + unsigned long len); + /* * Clear the page's PageMlocked(). This can be useful in a situation where * we want to unconditionally remove a page from the pagecache -- e.g., diff --git a/mm/mmap.c b/mm/mmap.c index 0584e540246e..81f5595a8490 100644 --- a/mm/mmap.c +++ b/mm/mmap.c @@ -1352,9 +1352,8 @@ static inline unsigned long round_hint_to_min(unsigned long hint) return hint; } -static inline int mlock_future_check(struct mm_struct *mm, - unsigned long flags, - unsigned long len) +int mlock_future_check(struct mm_struct *mm, unsigned long flags, + unsigned long len) { unsigned long locked, lock_limit;