From patchwork Fri Sep 28 05:39:46 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: john.hubbard@gmail.com X-Patchwork-Id: 10618937 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id CCCF3175A for ; Fri, 28 Sep 2018 05:40:02 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id B886F2B14F for ; Fri, 28 Sep 2018 05:40:02 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id ABE142B192; Fri, 28 Sep 2018 05:40:02 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-3.0 required=2.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,FREEMAIL_FROM,MAILING_LIST_MULTI,RCVD_IN_DNSWL_NONE autolearn=ham version=3.3.1 Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 68CFE2B177 for ; Fri, 28 Sep 2018 05:40:01 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id CF6EF8E0004; Fri, 28 Sep 2018 01:39:59 -0400 (EDT) Delivered-To: linux-mm-outgoing@kvack.org Received: by kanga.kvack.org (Postfix, from userid 40) id CA63A8E0001; Fri, 28 Sep 2018 01:39:59 -0400 (EDT) X-Original-To: int-list-linux-mm@kvack.org X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id B70248E0004; Fri, 28 Sep 2018 01:39:59 -0400 (EDT) X-Original-To: linux-mm@kvack.org X-Delivered-To: linux-mm@kvack.org Received: from mail-pf1-f197.google.com (mail-pf1-f197.google.com [209.85.210.197]) by kanga.kvack.org (Postfix) with ESMTP id 6505C8E0001 for ; Fri, 28 Sep 2018 01:39:59 -0400 (EDT) Received: by mail-pf1-f197.google.com with SMTP id f89-v6so5577848pff.7 for ; Thu, 27 Sep 2018 22:39:59 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:dkim-signature:from:to:cc:subject:date :message-id:in-reply-to:references:mime-version :content-transfer-encoding; bh=FVIsmZsD2UYUIajD9wCWvufUKrPcgTlr9PeHswP0yEs=; b=rXZxRuSRXym38NOJXssSMa2snhHg5iSihgym5RYGUSVaLZw/f/4lFZC+wu5BoMiWYd WsKH7tfoZGacoPO2VU/lcCXu/svQpEKU+UpcVyDmSeVqTO355Pmsv7qnb43zql0u8sxR 01UrHD1tvwDGGj5QAoJnBytyzLBWxlbrxaCTTbVINI4B+6JUfiWBrtvHtkBGXeKTwhNX XVbG5LrF6oU8PnwsSK3awWN0rd7wQ1DtVS/dFaFdWPSWEQhpu4n02rOM8og10Tqvzvd0 xCUdBUBKGff2BXp7n8+Dz1hW0k8oL8NHYwZ1dWu7uxO9qFHSxeriMXjClxGGewPU/Uzp EYow== X-Gm-Message-State: ABuFfohoUNE75cW3y2vvh7V1djKkFb/HIAdEjcBOxHRAdAiTCy7lyajV kFC9GtDyFwacm2QbDamyDGr3ErWMWOB4GnxWWloBqvDL3GbyZbaZoocQrr0hPYIc7pIR0bzvr8M 7l4IDPZlK23o4WvSxXnjyvvldayU7COkvbOhhknfPpSAu3oFYYNDAk5hmE/NXT76tKIphTURmx3 lD/fB6Fc1GOmlrNJaQz3Ol506eXZ/sTYnls/fFFxZz64E0AtZyViaPfzz3OomBuUcvsvnkXdyFV d8vpDwC5aABLt8qthUIleUm95YCWZO1oJoKaJFfVcMZgECn6gHaCGt04TbF2q8mzjYDmPKjuVey nePvB0J6nR5rYl95eNB/VJpGvGlpXa2LjbegSno5vr5VAvl8WFm3HcHVkwsq5ZF39yukq6zpFnU s X-Received: by 2002:a62:2315:: with SMTP id j21-v6mr608384pfj.90.1538113199102; Thu, 27 Sep 2018 22:39:59 -0700 (PDT) X-Received: by 2002:a62:2315:: with SMTP id j21-v6mr608324pfj.90.1538113198207; Thu, 27 Sep 2018 22:39:58 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1538113198; cv=none; d=google.com; s=arc-20160816; b=u1h/YKwJeuZlHqQnC27mJbvtnGo9o1JHkbmmC1k7h1jNXd1ab+t8EEp9h6ym4ZVjtW 39kABkc+q6lIJ7qdkIcFtvyyJGy992z4yKotLYhmG1H6Yec6rv+WvPk4OclpYbCQNKjV aDEhQciuZBpBZ7Yokir8Jm0nrKgUx5IcfiOt5xm0sRaxsAqV3LcE2uX5Xe7/T8xIQwKL tUACSTA6aoL6V2P9CMp1+MkG8X72NEJv1Qw2HzkMGWd/1wTUrf1EczLlc2KnYOEwq8d2 VwTursctVBDhyBuBoALg2CxsWoAHsanBWerRKRsrJ2PC1muUjweJT7o2AZFIMmNBJOiO HgfA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:dkim-signature; bh=FVIsmZsD2UYUIajD9wCWvufUKrPcgTlr9PeHswP0yEs=; b=hXN/vsO/1tZQrPDTbA3b+8l3qsFzOUmh5w3fP3grvMlJhdo1ddj2x2KUB+LLPdnxRV 2k5CLPEQYREEyKitVun8Qca7iCUP+1J30rdrIkNSzMoTdvoSSMKaiTBqRKbfZ1qd8n3Q 3Bz/67vwdYWdDlxcT6XsSx20bo6bfw4ZAfVCBUfA2gVYjRFA5nnuwHs5Squ4EWxN9u81 Hn0zytHr9SytE2FH9l0uA1C9hiSqDeLvOzT5EQ49sv7vHV1D8Z9peiKe2Y9PF/WARi/3 SiQkAQADEv778UCV9GPamxkjlx+46DmwR+aZdUkYA/0QbHfJWi48ePIq/jS8pDCHHiDG 8mOQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=n5r6tBaC; spf=pass (google.com: domain of john.hubbard@gmail.com designates 209.85.220.65 as permitted sender) smtp.mailfrom=john.hubbard@gmail.com; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: from mail-sor-f65.google.com (mail-sor-f65.google.com. [209.85.220.65]) by mx.google.com with SMTPS id t12-v6sor1121626pga.201.2018.09.27.22.39.58 for (Google Transport Security); Thu, 27 Sep 2018 22:39:58 -0700 (PDT) Received-SPF: pass (google.com: domain of john.hubbard@gmail.com designates 209.85.220.65 as permitted sender) client-ip=209.85.220.65; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=n5r6tBaC; spf=pass (google.com: domain of john.hubbard@gmail.com designates 209.85.220.65 as permitted sender) smtp.mailfrom=john.hubbard@gmail.com; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=FVIsmZsD2UYUIajD9wCWvufUKrPcgTlr9PeHswP0yEs=; b=n5r6tBaCfMwQJkMX1wNstsmX0XTetJz0PQnc4I6TTdGn5A0pBCYos1RpuFTQmM5hee ePBsKuqJClD80pR5uM+cOaCDaDiobtXHd4BugZ9R1lWYdqV3kMb/O1cbQx2j6jifnfPC OHYQKA1Z4f1/VkcgCqMJNMdxpDWpZCbVSPHwavstfckQEu0JsDhU4QTn82HhYQaLSXNS ektAk+Q6/Mat3WtnRAVZRR08nDBFtuDcbTkGsWB7Ps0t2T+XjWP3iYeWCvxEHdDTJ28d nP3DII/FHRK+Rc+fskScXoB3LUkO6/aUlAvFZvg/5NPsorwoQLAdV3O/w5WAONvGXDG/ ylsg== X-Google-Smtp-Source: ACcGV63VBR+76bN61j/R5pu3ZYOFO30IL/SSRAT4nbQ6f5aCW7Pi+BLOjTjb+ceaQUz6DgMOBj4qfg== X-Received: by 2002:a62:12c9:: with SMTP id 70-v6mr14833644pfs.140.1538113197825; Thu, 27 Sep 2018 22:39:57 -0700 (PDT) Received: from blueforge.nvidia.com (searspoint.nvidia.com. [216.228.112.21]) by smtp.gmail.com with ESMTPSA id u9-v6sm6569953pfi.104.2018.09.27.22.39.56 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Thu, 27 Sep 2018 22:39:56 -0700 (PDT) From: john.hubbard@gmail.com X-Google-Original-From: jhubbard@nvidia.com To: Matthew Wilcox , Michal Hocko , Christopher Lameter , Jason Gunthorpe , Dan Williams , Jan Kara , Al Viro Cc: linux-mm@kvack.org, LKML , linux-rdma , linux-fsdevel@vger.kernel.org, John Hubbard Subject: [PATCH 1/4] mm: get_user_pages: consolidate error handling Date: Thu, 27 Sep 2018 22:39:46 -0700 Message-Id: <20180928053949.5381-2-jhubbard@nvidia.com> X-Mailer: git-send-email 2.19.0 In-Reply-To: <20180928053949.5381-1-jhubbard@nvidia.com> References: <20180928053949.5381-1-jhubbard@nvidia.com> MIME-Version: 1.0 X-NVConfidentiality: public X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: X-Virus-Scanned: ClamAV using ClamSMTP From: John Hubbard An upcoming patch requires a way to operate on each page that any of the get_user_pages_*() variants returns. In preparation for that, consolidate the error handling for __get_user_pages(). This provides a single location (the "out:" label) for operating on the collected set of pages that are about to be returned. As long every use of the "ret" variable is being edited, rename "ret" --> "err", so that its name matches its true role. This also gets rid of two shadowed variable declarations, as a tiny beneficial a side effect. Reviewed-by: Jan Kara Signed-off-by: John Hubbard --- mm/gup.c | 37 ++++++++++++++++++++++--------------- 1 file changed, 22 insertions(+), 15 deletions(-) diff --git a/mm/gup.c b/mm/gup.c index 1abc8b4afff6..05ee7c18e59a 100644 --- a/mm/gup.c +++ b/mm/gup.c @@ -660,6 +660,7 @@ static long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, struct vm_area_struct **vmas, int *nonblocking) { long i = 0; + int err = 0; unsigned int page_mask; struct vm_area_struct *vma = NULL; @@ -685,18 +686,19 @@ static long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, if (!vma || start >= vma->vm_end) { vma = find_extend_vma(mm, start); if (!vma && in_gate_area(mm, start)) { - int ret; - ret = get_gate_page(mm, start & PAGE_MASK, + err = get_gate_page(mm, start & PAGE_MASK, gup_flags, &vma, pages ? &pages[i] : NULL); - if (ret) - return i ? : ret; + if (err) + goto out; page_mask = 0; goto next_page; } - if (!vma || check_vma_flags(vma, gup_flags)) - return i ? : -EFAULT; + if (!vma || check_vma_flags(vma, gup_flags)) { + err = -EFAULT; + goto out; + } if (is_vm_hugetlb_page(vma)) { i = follow_hugetlb_page(mm, vma, pages, vmas, &start, &nr_pages, i, @@ -709,23 +711,25 @@ static long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, * If we have a pending SIGKILL, don't keep faulting pages and * potentially allocating memory. */ - if (unlikely(fatal_signal_pending(current))) - return i ? i : -ERESTARTSYS; + if (unlikely(fatal_signal_pending(current))) { + err = -ERESTARTSYS; + goto out; + } cond_resched(); page = follow_page_mask(vma, start, foll_flags, &page_mask); if (!page) { - int ret; - ret = faultin_page(tsk, vma, start, &foll_flags, + err = faultin_page(tsk, vma, start, &foll_flags, nonblocking); - switch (ret) { + switch (err) { case 0: goto retry; case -EFAULT: case -ENOMEM: case -EHWPOISON: - return i ? i : ret; + goto out; case -EBUSY: - return i; + err = 0; + goto out; case -ENOENT: goto next_page; } @@ -737,7 +741,8 @@ static long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, */ goto next_page; } else if (IS_ERR(page)) { - return i ? i : PTR_ERR(page); + err = PTR_ERR(page); + goto out; } if (pages) { pages[i] = page; @@ -757,7 +762,9 @@ static long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, start += page_increm * PAGE_SIZE; nr_pages -= page_increm; } while (nr_pages); - return i; + +out: + return i ? i : err; } static bool vma_permits_fault(struct vm_area_struct *vma,