From patchwork Tue Aug 18 03:34:04 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Nicholas Piggin X-Patchwork-Id: 11719759 Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 3DC5B722 for ; Tue, 18 Aug 2020 03:34:18 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id F188F2076D for ; Tue, 18 Aug 2020 03:34:17 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=fail reason="signature verification failed" (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b="JTsxTQCW" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org F188F2076D Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=gmail.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 0D73A6B0002; Mon, 17 Aug 2020 23:34:17 -0400 (EDT) Delivered-To: linux-mm-outgoing@kvack.org Received: by kanga.kvack.org (Postfix, from userid 40) id 05F5E6B0005; Mon, 17 Aug 2020 23:34:16 -0400 (EDT) X-Original-To: int-list-linux-mm@kvack.org X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id E42256B0006; Mon, 17 Aug 2020 23:34:16 -0400 (EDT) X-Original-To: linux-mm@kvack.org X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0012.hostedemail.com [216.40.44.12]) by kanga.kvack.org (Postfix) with ESMTP id CAB846B0002 for ; Mon, 17 Aug 2020 23:34:16 -0400 (EDT) Received: from smtpin05.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay04.hostedemail.com (Postfix) with ESMTP id 88D2E2470 for ; Tue, 18 Aug 2020 03:34:16 +0000 (UTC) X-FDA: 77162271312.05.beast85_2008ca82701c Received: from filter.hostedemail.com (10.5.16.251.rfc1918.com [10.5.16.251]) by smtpin05.hostedemail.com (Postfix) with ESMTP id 61EB518017807 for ; Tue, 18 Aug 2020 03:34:16 +0000 (UTC) X-Spam-Summary: 1,0,0,1bbd12d9004457ab,d41d8cd98f00b204,npiggin@gmail.com,,RULES_HIT:2:41:69:355:379:541:800:960:968:973:988:989:1260:1311:1314:1345:1437:1515:1535:1605:1606:1730:1747:1777:1792:2393:2559:2562:2693:2895:3138:3139:3140:3141:3142:3165:3865:3866:3867:3868:3870:3871:3872:3873:3874:4037:4119:4250:4321:4362:4605:5007:6119:6120:6261:6653:7514:7903:8660:9008:9010:9413:9592:10004:11026:11473:11658:11914:12043:12296:12297:12438:12517:12519:12555:12683:12895:13148:13161:13229:13230:13894:14096:14687:21080:21433:21444:21451:21627:21666:21740:21795:21809:21939:21972:21990:30003:30012:30034:30045:30051:30054:30056:30070:30075,0,RBL:209.85.214.195:@gmail.com:.lbl8.mailshell.net-66.100.201.100 62.50.0.100;04yrzt7n1yp4jez156e9je18bdryuycy93auxsoacrs6okwrqydwidkumyhdnfb.mxg3qdr818a37jhuzjnzq34tywjx1qinjyft3pa5we54e7sbbbsg9c75ztqqoi7.h-lbl8.mailshell.net-223.238.255.100,CacheIP:none,Bayesian:0.5,0.5,0.5,Netcheck:none,DomainCache:0,MSF:not bulk,SPF:fp,MSBL:0,DNSBL:neutral,Custom_ rules:0: X-HE-Tag: beast85_2008ca82701c X-Filterd-Recvd-Size: 8928 Received: from mail-pl1-f195.google.com (mail-pl1-f195.google.com [209.85.214.195]) by imf02.hostedemail.com (Postfix) with ESMTP for ; Tue, 18 Aug 2020 03:34:15 +0000 (UTC) Received: by mail-pl1-f195.google.com with SMTP id f5so8554447plr.9 for ; Mon, 17 Aug 2020 20:34:15 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=kXKR3cXpwpt0q9jgFz26UJPnEar0RH5gNFSzQ9D4WUQ=; b=JTsxTQCWOkw6T9c3YtWdB1NVIAooJ/0TYLt2HeP3tRrpo9IrPH5fcSJC/W1TF6hUrK 36Q88tCA5xip+UxdciaZbVfDYgfeyDGUIlbtktjwoxbZTeYrUnbZdG9gFRAnIwKqJwdP CmGqb7y9E+ke6lvgFU5x9I3iM9wDvEHVONeLH5m/QC2OyjTuFzQr+1ki+1vWIx4c71RY cYA4y4P1JW8KO+6gnUICfuT2g5W25/HCbpSLRvLO85Mdpzi+D1wSk/sEAtEssZ3j+vfB n34t1a9YyqYAUV0FR8/FV1KruetVFbpMilo3/y6737GJKAD+w5XJS50Ge4nPNk/NYv2B 7DIQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=kXKR3cXpwpt0q9jgFz26UJPnEar0RH5gNFSzQ9D4WUQ=; b=psBSj1OpbLH2G5ZzpngWgTuEJ5YEJaL0HmjAlIA7Ps5AmGFs1RBYtQa04A1dle89xd yAwvTNm1LL3444m1ZkJXbow7K5MfTlmmWzfkCvU7v/ccbqhkzYTym5vj1iDVc5ZL0IZ2 tRabmYWdrcDLa4Un/btDKbFoKB/d8SCtepPrV77NgXd2OKUxluulq6dtilNo3bZpZErB Kv59/cFbb9DqDcQ079/w1sdTIONjqrurEuEyuGcCyzi/2Wys/fusDX3NcWrXmOBfdQBl jWSi8r0VeiRml0X3rRiNKvE45hHZnT1OM9mX7KG86EgRSf00uqhJckIHI5LRTIM+ABHM rBBQ== X-Gm-Message-State: AOAM530x/BZ3/HXvmyd+Kl1SGrRxyfWgFtUIilvxWLGWeQ+sT0z8wzBv Cx5DOl1k9tzf2LDvur3r5FI= X-Google-Smtp-Source: ABdhPJyW1IrlExwvUFobL+ijBXLtHQla2kejcW1bkPDfbMej7nAW5/8WB8glAdU0v5nRmt+I6vLSLA== X-Received: by 2002:a17:90b:349:: with SMTP id fh9mr14332259pjb.73.1597721654856; Mon, 17 Aug 2020 20:34:14 -0700 (PDT) Received: from bobo.ozlabs.ibm.com (193-116-193-175.tpgi.com.au. [193.116.193.175]) by smtp.gmail.com with ESMTPSA id b18sm19823352pgj.12.2020.08.17.20.34.11 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 17 Aug 2020 20:34:14 -0700 (PDT) From: Nicholas Piggin To: "David S. Miller" Cc: Nicholas Piggin , Jens Axboe , "Aneesh Kumar K.V" , linux-mm@kvack.org, linuxppc-dev@lists.ozlabs.org, linux-kernel@vger.kernel.org, sparclinux@vger.kernel.org Subject: [RFC PATCH 1/2] sparc64: remove mm_cpumask clearing to fix kthread_use_mm race Date: Tue, 18 Aug 2020 13:34:04 +1000 Message-Id: <20200818033405.170457-1-npiggin@gmail.com> X-Mailer: git-send-email 2.23.0 MIME-Version: 1.0 X-Rspamd-Queue-Id: 61EB518017807 X-Spamd-Result: default: False [0.00 / 100.00] X-Rspamd-Server: rspam04 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: The de facto (and apparently uncommented) standard for using an mm had, thanks to this code in sparc if nothing else, been that you must have a reference on mm_users *and that reference must have been obtained with mmget()*, i.e., from a thread with a reference to mm_users that had used the mm. The introduction of mmget_not_zero() in commit d2005e3f41d4 ("userfaultfd: don't pin the user memory in userfaultfd_file_create()") allowed mm_count holders to aoperate on user mappings asynchronously from the actual threads using the mm, but they were not to load those mappings into their TLB (i.e., walking vmas and page tables is okay, kthread_use_mm() is not). io_uring 2b188cc1bb857 ("Add io_uring IO interface") added code which does a kthread_use_mm() from a mmget_not_zero() refcount. The problem with this is code which previously assumed mm == current->mm and mm->mm_users == 1 implies the mm will remain single-threaded at least until this thread creates another mm_users reference, has now broken. arch/sparc/kernel/smp_64.c: if (atomic_read(&mm->mm_users) == 1) { cpumask_copy(mm_cpumask(mm), cpumask_of(cpu)); goto local_flush_and_out; } vs fs/io_uring.c if (unlikely(!(ctx->flags & IORING_SETUP_SQPOLL) || !mmget_not_zero(ctx->sqo_mm))) return -EFAULT; kthread_use_mm(ctx->sqo_mm); mmget_not_zero() could come in right after the mm_users == 1 test, then kthread_use_mm() which sets its CPU in the mm_cpumask. That update could be lost if cpumask_copy() occurs afterward. I propose we fix this by allowing mmget_not_zero() to be a first-class reference, and not have this obscure undocumented and unchecked restriction. The basic fix for sparc64 is to remove its mm_cpumask clearing code. The optimisation could be effectively restored by sending IPIs to mm_cpumask members and having them remove themselves from mm_cpumask. This is more tricky so I leave it as an exercise for someone with a sparc64 SMP. powerpc has a (currently similarly broken) example. not-yet-Signed-off-by: Nicholas Piggin --- arch/sparc/kernel/smp_64.c | 58 ++++++-------------------------------- 1 file changed, 9 insertions(+), 49 deletions(-) diff --git a/arch/sparc/kernel/smp_64.c b/arch/sparc/kernel/smp_64.c index e286e2badc8a..41423c6077d0 100644 --- a/arch/sparc/kernel/smp_64.c +++ b/arch/sparc/kernel/smp_64.c @@ -1039,38 +1039,9 @@ void smp_fetch_global_pmu(void) * are flush_tlb_*() routines, and these run after flush_cache_*() * which performs the flushw. * - * The SMP TLB coherency scheme we use works as follows: - * - * 1) mm->cpu_vm_mask is a bit mask of which cpus an address - * space has (potentially) executed on, this is the heuristic - * we use to avoid doing cross calls. - * - * Also, for flushing from kswapd and also for clones, we - * use cpu_vm_mask as the list of cpus to make run the TLB. - * - * 2) TLB context numbers are shared globally across all processors - * in the system, this allows us to play several games to avoid - * cross calls. - * - * One invariant is that when a cpu switches to a process, and - * that processes tsk->active_mm->cpu_vm_mask does not have the - * current cpu's bit set, that tlb context is flushed locally. - * - * If the address space is non-shared (ie. mm->count == 1) we avoid - * cross calls when we want to flush the currently running process's - * tlb state. This is done by clearing all cpu bits except the current - * processor's in current->mm->cpu_vm_mask and performing the - * flush locally only. This will force any subsequent cpus which run - * this task to flush the context from the local tlb if the process - * migrates to another cpu (again). - * - * 3) For shared address spaces (threads) and swapping we bite the - * bullet for most cases and perform the cross call (but only to - * the cpus listed in cpu_vm_mask). - * - * The performance gain from "optimizing" away the cross call for threads is - * questionable (in theory the big win for threads is the massive sharing of - * address space state across processors). + * mm->cpu_vm_mask is a bit mask of which cpus an address + * space has (potentially) executed on, this is the heuristic + * we use to limit cross calls. */ /* This currently is only used by the hugetlb arch pre-fault @@ -1082,16 +1053,10 @@ void smp_flush_tlb_mm(struct mm_struct *mm) u32 ctx = CTX_HWBITS(mm->context); int cpu = get_cpu(); - if (atomic_read(&mm->mm_users) == 1) { - cpumask_copy(mm_cpumask(mm), cpumask_of(cpu)); - goto local_flush_and_out; - } - smp_cross_call_masked(&xcall_flush_tlb_mm, ctx, 0, 0, mm_cpumask(mm)); -local_flush_and_out: __flush_tlb_mm(ctx, SECONDARY_CONTEXT); put_cpu(); @@ -1120,11 +1085,8 @@ void smp_flush_tlb_pending(struct mm_struct *mm, unsigned long nr, unsigned long info.nr = nr; info.vaddrs = vaddrs; - if (mm == current->mm && atomic_read(&mm->mm_users) == 1) - cpumask_copy(mm_cpumask(mm), cpumask_of(cpu)); - else - smp_call_function_many(mm_cpumask(mm), tlb_pending_func, - &info, 1); + smp_call_function_many(mm_cpumask(mm), tlb_pending_func, + &info, 1); __flush_tlb_pending(ctx, nr, vaddrs); @@ -1136,12 +1098,10 @@ void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr) unsigned long context = CTX_HWBITS(mm->context); int cpu = get_cpu(); - if (mm == current->mm && atomic_read(&mm->mm_users) == 1) - cpumask_copy(mm_cpumask(mm), cpumask_of(cpu)); - else - smp_cross_call_masked(&xcall_flush_tlb_page, - context, vaddr, 0, - mm_cpumask(mm)); + smp_cross_call_masked(&xcall_flush_tlb_page, + context, vaddr, 0, + mm_cpumask(mm)); + __flush_tlb_page(context, vaddr); put_cpu();