Message ID | 20191211235108.GA85068@google.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | rmap for shmem pages | expand |
On Wed, Dec 11, 2019 at 06:51:08PM -0500, Joel Fernandes wrote: > Hi, > I am looking into a weird behavior of the VmRSS counter in /proc/pid/status > which I think is a bug. > > The issue is if a memfd is mapped into the same process twice as MAP_SHARED, and > you write to both maps, then the RSS doubles even though the pages were > mapped elsewhere within the same address space before. We expect RSS to > increase only once in this case, not twice. VmRSS is a best effort. Meeting your expectation would require rmap walk (as your patch does). It's way too expensive for fault path (especially under page table lock) just to keep percise statistics. Sorry, but you will not get it.
On Thu, Dec 12, 2019 at 01:55:21PM +0300, Kirill A. Shutemov wrote: > On Wed, Dec 11, 2019 at 06:51:08PM -0500, Joel Fernandes wrote: > > Hi, > > I am looking into a weird behavior of the VmRSS counter in /proc/pid/status > > which I think is a bug. > > > > The issue is if a memfd is mapped into the same process twice as MAP_SHARED, and > > you write to both maps, then the RSS doubles even though the pages were > > mapped elsewhere within the same address space before. We expect RSS to > > increase only once in this case, not twice. > > VmRSS is a best effort. Meeting your expectation would require rmap walk > (as your patch does). It's way too expensive for fault path (especially > under page table lock) just to keep percise statistics. Sorry, but you > will not get it. No problem. I was under the same impression that this is difficult to do in a scalable way. I am glad I reached out but if anyone has any other ideas, let me know. thanks, - Joel > > -- > Kirill A. Shutemov
diff --git a/mm/memory.c b/mm/memory.c index 606da187d1de..4fca5d7e9d66 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -3360,6 +3361,8 @@ static vm_fault_t do_set_pmd(struct vm_fault *vmf, struct page *page) * * Return: %0 on success, %VM_FAULT_ code in case of error. */ +bool page_has_single_mm(struct page *page, struct mm_struct *mm); + vm_fault_t alloc_set_pte(struct vm_fault *vmf, struct mem_cgroup *memcg, struct page *page) { @@ -3399,8 +3402,14 @@ vm_fault_t alloc_set_pte(struct vm_fault *vmf, struct mem_cgroup *memcg, mem_cgroup_commit_charge(page, memcg, false, false); lru_cache_add_active_or_unevictable(page, vma); } else { - inc_mm_counter_fast(vma->vm_mm, mm_counter_file(page)); page_add_file_rmap(page, false); + + if (page_has_single_mm(page, vma->vm_mm)) { + trace_printk("inc 2 mm_counter_file counter=%d\n", mm_counter_file(page)); + inc_mm_counter_fast(vma->vm_mm, mm_counter_file(page)); + } else { + trace_printk("inc 2 mm_counter_file counter=%d skipped\n", mm_counter_file(page)); + } } set_pte_at(vma->vm_mm, vmf->address, vmf->pte, entry); diff --git a/mm/rmap.c b/mm/rmap.c index b3e381919835..580da8469f0d 100644 --- a/mm/rmap.c +++ b/mm/rmap.c @@ -1273,6 +1273,39 @@ static void page_remove_file_rmap(struct page *page, bool compound) unlock_page_memcg(page); } +struct page_single_mm_ctx { + struct mm_struct *mm; + bool single; +}; + +static bool __page_has_single_mm(struct page *page, struct vm_area_struct *vma, + unsigned long addr, void *ctx_in) +{ + struct page_single_mm_ctx *ctx = ctx_in; + + if (vma->vm_mm == ctx->mm) + return true; + + ctx->single = false; + return false; +} + +bool page_has_single_mm(struct page *page, struct mm_struct *mm) +{ + struct page_single_mm_ctx ctx; + + ctx.mm = mm; + ctx.single = true; + + struct rmap_walk_control rwc = { + .rmap_one = __page_has_single_mm, + .arg = &ctx, + }; + + rmap_walk(page, &rwc); + return ctx.single; +} + static void page_remove_anon_compound_rmap(struct page *page) { int i, nr;