From patchwork Thu May 15 15:56:52 2014 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Weston Andros Adamson X-Patchwork-Id: 4183471 Return-Path: X-Original-To: patchwork-linux-nfs@patchwork.kernel.org Delivered-To: patchwork-parsemail@patchwork2.web.kernel.org Received: from mail.kernel.org (mail.kernel.org [198.145.19.201]) by patchwork2.web.kernel.org (Postfix) with ESMTP id CDB94C0ACC for ; Thu, 15 May 2014 15:57:35 +0000 (UTC) Received: from mail.kernel.org (localhost [127.0.0.1]) by mail.kernel.org (Postfix) with ESMTP id D5B96201FE for ; Thu, 15 May 2014 15:57:34 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id E1F4F20379 for ; Thu, 15 May 2014 15:57:33 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1755189AbaEOP5a (ORCPT ); Thu, 15 May 2014 11:57:30 -0400 Received: from mail-ie0-f170.google.com ([209.85.223.170]:47237 "EHLO mail-ie0-f170.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755172AbaEOP53 (ORCPT ); Thu, 15 May 2014 11:57:29 -0400 Received: by mail-ie0-f170.google.com with SMTP id ar20so1210888iec.15 for ; Thu, 15 May 2014 08:57:28 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=nhJg9QrhDzO+MVYXtlvPugJYsCrZ+3gEu1xyU06p9/c=; b=QF1ltgOM5avu6qz11tvMpTXIG/BLeRMfzkd6n0VQ40UMRsnqoGgEk0HK1puAAqymoJ vBgJWICY0YOP3+hB8PcDHpyvdtoRaZJFGjaIq9BnKdzyyclXdakCbP0cXnEvqVDA/+nI Y58T82OX/l27TZc2vW30vyn3U5mjk0ojvm+HacYy98N1Pk0RBLQhWPbCPF26A0Mhy2xO CpPzxa79XaW0GMOwvxH5hknrc4AtBYaVt8m7Ntdc9nS5FpTOw4t3/sCQ6QWzBAuPnoP7 TXTYZFbGG1mn51avHL32wmhFfbNXirRFs9Te5bKpbX8v78DOIAMYnYG0xyBhsJTyg5NG mhDw== X-Gm-Message-State: ALoCoQnyzEAqRtv/DJnzJF+wykleX4xyfukTpQkvO5khWFGhYbS1okntL50eNeGu8USyOUPb7n+q X-Received: by 10.50.129.97 with SMTP id nv1mr74116056igb.32.1400169448520; Thu, 15 May 2014 08:57:28 -0700 (PDT) Received: from gavrio-wifi.robotsandstuff.fake (c-98-209-19-144.hsd1.mi.comcast.net. [98.209.19.144]) by mx.google.com with ESMTPSA id jf5sm13964856igb.19.2014.05.15.08.57.26 for (version=TLSv1 cipher=RC4-SHA bits=128/128); Thu, 15 May 2014 08:57:27 -0700 (PDT) From: Weston Andros Adamson To: trond.myklebust@primarydata.com Cc: linux-nfs@vger.kernel.org, Weston Andros Adamson Subject: [PATCH v3 13/18] nfs: use > 1 request to handle bsize < PAGE_SIZE Date: Thu, 15 May 2014 11:56:52 -0400 Message-Id: <1400169417-28245-14-git-send-email-dros@primarydata.com> X-Mailer: git-send-email 1.8.5.2 (Apple Git-48) In-Reply-To: <1400169417-28245-1-git-send-email-dros@primarydata.com> References: <1400169417-28245-1-git-send-email-dros@primarydata.com> Sender: linux-nfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org X-Spam-Status: No, score=-7.5 required=5.0 tests=BAYES_00, RCVD_IN_DNSWL_HI, RP_MATCHES_RCVD, UNPARSEABLE_RELAY autolearn=ham version=3.3.1 X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on mail.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP Use the newly added support for multiple requests per page for rsize/wsize < PAGE_SIZE, instead of having multiple read / write data structures per pageio header. This allows us to get rid of nfs_pgio_multi. Signed-off-by: Weston Andros Adamson --- fs/nfs/pagelist.c | 80 ++++++++----------------------------------------------- 1 file changed, 11 insertions(+), 69 deletions(-) diff --git a/fs/nfs/pagelist.c b/fs/nfs/pagelist.c index 44aef5a..0871a95 100644 --- a/fs/nfs/pagelist.c +++ b/fs/nfs/pagelist.c @@ -443,21 +443,13 @@ nfs_wait_on_request(struct nfs_page *req) size_t nfs_generic_pg_test(struct nfs_pageio_descriptor *desc, struct nfs_page *prev, struct nfs_page *req) { - if (!prev) - return req->wb_bytes; - /* - * FIXME: ideally we should be able to coalesce all requests - * that are not block boundary aligned, but currently this - * is problematic for the case of bsize < PAGE_CACHE_SIZE, - * since nfs_flush_multi and nfs_pagein_multi assume you - * can have only one struct nfs_page. - */ - if (desc->pg_bsize < PAGE_SIZE) + if (desc->pg_count > desc->pg_bsize) { + /* should never happen */ + WARN_ON_ONCE(1); return 0; + } - if (desc->pg_count + req->wb_bytes <= desc->pg_bsize) - return req->wb_bytes; - return 0; + return min(desc->pg_bsize - desc->pg_count, (size_t)req->wb_bytes); } EXPORT_SYMBOL_GPL(nfs_generic_pg_test); @@ -767,50 +759,6 @@ static void nfs_pgio_result(struct rpc_task *task, void *calldata) } /* - * Generate multiple small requests to read or write a single - * contiguous dirty on one page. - */ -static int nfs_pgio_multi(struct nfs_pageio_descriptor *desc, - struct nfs_pgio_header *hdr) -{ - struct nfs_page *req = hdr->req; - struct page *page = req->wb_page; - struct nfs_pgio_data *data; - size_t wsize = desc->pg_bsize, nbytes; - unsigned int offset; - int requests = 0; - struct nfs_commit_info cinfo; - - nfs_init_cinfo(&cinfo, desc->pg_inode, desc->pg_dreq); - - if ((desc->pg_ioflags & FLUSH_COND_STABLE) && - (desc->pg_moreio || nfs_reqs_to_commit(&cinfo) || - desc->pg_count > wsize)) - desc->pg_ioflags &= ~FLUSH_COND_STABLE; - - offset = 0; - nbytes = desc->pg_count; - do { - size_t len = min(nbytes, wsize); - - data = nfs_pgio_data_alloc(hdr, 1); - if (!data) - return nfs_pgio_error(desc, hdr); - data->pages.pagevec[0] = page; - nfs_pgio_rpcsetup(data, len, offset, desc->pg_ioflags, &cinfo); - list_add(&data->list, &hdr->rpc_list); - requests++; - nbytes -= len; - offset += len; - } while (nbytes != 0); - - nfs_list_remove_request(req); - nfs_list_add_request(req, &hdr->pages); - desc->pg_rpc_callops = &nfs_pgio_common_ops; - return 0; -} - -/* * Create an RPC task for the given read or write request and kick it. * The page must have been locked by the caller. * @@ -818,8 +766,8 @@ static int nfs_pgio_multi(struct nfs_pageio_descriptor *desc, * This is the case if nfs_updatepage detects a conflicting request * that has been written but not committed. */ -static int nfs_pgio_one(struct nfs_pageio_descriptor *desc, - struct nfs_pgio_header *hdr) +int nfs_generic_pgio(struct nfs_pageio_descriptor *desc, + struct nfs_pgio_header *hdr) { struct nfs_page *req; struct page **pages; @@ -851,6 +799,7 @@ static int nfs_pgio_one(struct nfs_pageio_descriptor *desc, desc->pg_rpc_callops = &nfs_pgio_common_ops; return 0; } +EXPORT_SYMBOL_GPL(nfs_generic_pgio); static int nfs_generic_pg_pgios(struct nfs_pageio_descriptor *desc) { @@ -876,15 +825,6 @@ static int nfs_generic_pg_pgios(struct nfs_pageio_descriptor *desc) return ret; } -int nfs_generic_pgio(struct nfs_pageio_descriptor *desc, - struct nfs_pgio_header *hdr) -{ - if (desc->pg_bsize < PAGE_CACHE_SIZE) - return nfs_pgio_multi(desc, hdr); - return nfs_pgio_one(desc, hdr); -} -EXPORT_SYMBOL_GPL(nfs_generic_pgio); - static bool nfs_match_open_context(const struct nfs_open_context *ctx1, const struct nfs_open_context *ctx2) { @@ -926,7 +866,9 @@ static bool nfs_can_coalesce_requests(struct nfs_page *prev, return false; } size = pgio->pg_ops->pg_test(pgio, prev, req); - WARN_ON_ONCE(size && size != req->wb_bytes); + WARN_ON_ONCE(size > req->wb_bytes); + if (size && size < req->wb_bytes) + req->wb_bytes = size; return size > 0; }