From patchwork Mon May 27 16:36:08 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Christoph Hellwig X-Patchwork-Id: 13675644 Received: from bombadil.infradead.org (bombadil.infradead.org [198.137.202.133]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 16C76155C96; Mon, 27 May 2024 16:36:23 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=198.137.202.133 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1716827785; cv=none; b=V6Y68uJCsvRo6JZs4YRtClOSxQfraflnDsVARCEOaSkhPk2OC+9rueklfJdmpYSpH4Lu3MAIjulu2Xy3jB9P6eQbVlU6L/06C2uIapRKgsTtUUTZbZbB8tAfoBP+MC8NcOlgIOzqPTNLDbvqwHIAIWpplSOC5SfNsqF8EAqlsqQ= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1716827785; c=relaxed/simple; bh=5PXQDAmsoR4Mi2gYo4EP22yZCBVVSAfwuHdb3qiEX/M=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=oXPtghK3kUV6ndX/pVhUYp6J4j82/ozQFU1WngGbu7wf4THVoT16e8HI8MjGcoMyKRBUJVIoscKPwBM1Twtk75CeeYc4fJsPZSFHLFX/07Kl63wPxrFd1nEc7lyHrosdV0nUKTLAgzKUcmwTyEFlhVh01ksN+TgdJMJizvcyS1g= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=lst.de; spf=none smtp.mailfrom=bombadil.srs.infradead.org; dkim=pass (2048-bit key) header.d=infradead.org header.i=@infradead.org header.b=DF+0PB+J; arc=none smtp.client-ip=198.137.202.133 Authentication-Results: smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=lst.de Authentication-Results: smtp.subspace.kernel.org; spf=none smtp.mailfrom=bombadil.srs.infradead.org Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=infradead.org header.i=@infradead.org header.b="DF+0PB+J" DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=bombadil.20210309; h=Content-Transfer-Encoding: MIME-Version:References:In-Reply-To:Message-ID:Date:Subject:Cc:To:From:Sender :Reply-To:Content-Type:Content-ID:Content-Description; bh=27+lP32Q2mw19iey55M36pSytXlr5S4+SrngC3O6v+s=; b=DF+0PB+J1G3GjShibBUkkXKzbg fXBJp4C12z3fLlSUjFb8iSZ12AevUEdRfRjZ65AWcrNk697Svt6DuF5gVqg4QjbjnUClInBMEI1fO TfEbmNaMYRqU/2QXTmXXyJL+AvEdUbrZ+JSijhLQH8oOgsoRdGmy8m7OGPwOckiF+0Z6qGrT1GbQv T/4dxOFKBIJdGP/bWENwVdcqF0qb34xY2V4MCNg0VQwK49HqlXvDjpSaCsAvfsFMLwrrOFg7l6wSd dCI/QdL+yV5U3WF7sB3kYhStTLsR9aFpkkG+H5/jCwMl6/YKu7F1t0WFZn5pMml3EY9duNuMKrv16 rsI69DyQ==; Received: from 2a02-8389-2341-5b80-3177-e4c1-2108-f294.cable.dynamic.v6.surfer.at ([2a02:8389:2341:5b80:3177:e4c1:2108:f294] helo=localhost) by bombadil.infradead.org with esmtpsa (Exim 4.97.1 #2 (Red Hat Linux)) id 1sBdKh-0000000Ft6T-11pI; Mon, 27 May 2024 16:36:23 +0000 From: Christoph Hellwig To: Trond Myklebust , Anna Schumaker , Matthew Wilcox Cc: linux-nfs@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org Subject: [PATCH 1/2] filemap: Convert generic_perform_write() to support large folios Date: Mon, 27 May 2024 18:36:08 +0200 Message-ID: <20240527163616.1135968-2-hch@lst.de> X-Mailer: git-send-email 2.43.0 In-Reply-To: <20240527163616.1135968-1-hch@lst.de> References: <20240527163616.1135968-1-hch@lst.de> Precedence: bulk X-Mailing-List: linux-nfs@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 X-SRS-Rewrite: SMTP reverse-path rewritten from by bombadil.infradead.org. See http://www.infradead.org/rpr.html From: "Matthew Wilcox (Oracle)" Modelled after the loop in iomap_write_iter(), copy larger chunks from userspace if the filesystem has created large folios. Signed-off-by: Matthew Wilcox (Oracle) [hch: use mapping_max_folio_size to keep supporting file systems that do not support large folios] Signed-off-by: Christoph Hellwig Tested-by: Shaun Tancheff --- mm/filemap.c | 40 +++++++++++++++++++++++++--------------- 1 file changed, 25 insertions(+), 15 deletions(-) diff --git a/mm/filemap.c b/mm/filemap.c index 382c3d06bfb10c..860728e26ccf32 100644 --- a/mm/filemap.c +++ b/mm/filemap.c @@ -3981,21 +3981,24 @@ ssize_t generic_perform_write(struct kiocb *iocb, struct iov_iter *i) loff_t pos = iocb->ki_pos; struct address_space *mapping = file->f_mapping; const struct address_space_operations *a_ops = mapping->a_ops; + size_t chunk = mapping_max_folio_size(mapping); long status = 0; ssize_t written = 0; do { struct page *page; - unsigned long offset; /* Offset into pagecache page */ - unsigned long bytes; /* Bytes to write to page */ + struct folio *folio; + size_t offset; /* Offset into folio */ + size_t bytes; /* Bytes to write to folio */ size_t copied; /* Bytes copied from user */ void *fsdata = NULL; - offset = (pos & (PAGE_SIZE - 1)); - bytes = min_t(unsigned long, PAGE_SIZE - offset, - iov_iter_count(i)); + bytes = iov_iter_count(i); +retry: + offset = pos & (chunk - 1); + bytes = min(chunk - offset, bytes); + balance_dirty_pages_ratelimited(mapping); -again: /* * Bring in the user page that we will copy from _first_. * Otherwise there's a nasty deadlock on copying from the @@ -4017,11 +4020,16 @@ ssize_t generic_perform_write(struct kiocb *iocb, struct iov_iter *i) if (unlikely(status < 0)) break; + folio = page_folio(page); + offset = offset_in_folio(folio, pos); + if (bytes > folio_size(folio) - offset) + bytes = folio_size(folio) - offset; + if (mapping_writably_mapped(mapping)) - flush_dcache_page(page); + flush_dcache_folio(folio); - copied = copy_page_from_iter_atomic(page, offset, bytes, i); - flush_dcache_page(page); + copied = copy_folio_from_iter_atomic(folio, offset, bytes, i); + flush_dcache_folio(folio); status = a_ops->write_end(file, mapping, pos, bytes, copied, page, fsdata); @@ -4039,14 +4047,16 @@ ssize_t generic_perform_write(struct kiocb *iocb, struct iov_iter *i) * halfway through, might be a race with munmap, * might be severe memory pressure. */ - if (copied) + if (chunk > PAGE_SIZE) + chunk /= 2; + if (copied) { bytes = copied; - goto again; + goto retry; + } + } else { + pos += status; + written += status; } - pos += status; - written += status; - - balance_dirty_pages_ratelimited(mapping); } while (iov_iter_count(i)); if (!written) From patchwork Mon May 27 16:36:09 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Christoph Hellwig X-Patchwork-Id: 13675645 Received: from bombadil.infradead.org (bombadil.infradead.org [198.137.202.133]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id AF83D155C96; Mon, 27 May 2024 16:36:26 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=198.137.202.133 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1716827788; cv=none; b=Xbv3V1kLBGKa+R7qPrbOR7BiX4qKOwll3wHFadHdQC+x4jIU6I19CDztxcEkN0W/B6GjRWl2YVwRDLgTros+tc8v82ozi1nr5nb3RiCiNuFmNcT7M/Cx2QPiFOwX0qinxWNRZo87A3tILtR39JDSX+sI4gVc/8GIxZ1JBjL2O6o= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1716827788; c=relaxed/simple; bh=6XZlotLjP3r5f5t9P/KPPR+9SK+bBfdsrRTz57vSURc=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=UTkFXVGhfM7AfwxBSX8xhFZsQ4ukf+efjTCvTWWzWRbkeqSQCeitgduxhKPjgZzrMnKrUp6H+aB6CPi5CTjpyLAyrWkekMnwGJcPCT7Vp3sH2PfZnNkkOV3x5WbHRO493Nbt4wrhun4LUlpJVFVMILReq3JWQ9mt9/Ovkgn2IbA= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=lst.de; spf=none smtp.mailfrom=bombadil.srs.infradead.org; dkim=pass (2048-bit key) header.d=infradead.org header.i=@infradead.org header.b=K2OYVydg; arc=none smtp.client-ip=198.137.202.133 Authentication-Results: smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=lst.de Authentication-Results: smtp.subspace.kernel.org; spf=none smtp.mailfrom=bombadil.srs.infradead.org Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=infradead.org header.i=@infradead.org header.b="K2OYVydg" DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=bombadil.20210309; h=Content-Transfer-Encoding: MIME-Version:References:In-Reply-To:Message-ID:Date:Subject:Cc:To:From:Sender :Reply-To:Content-Type:Content-ID:Content-Description; bh=K4Mxq5Sr+5purWqI8GGeOoyAJxibrH1gh6qH/UJ9bOw=; b=K2OYVydgR+//Tm+fbwytqZykcF MrgUbGSku0QuXH1fWLVjNH2ND/9nYF9fqS34hhFdndgKYKYGc2T7RsAgcq628Wt96VO7vdDbnlTBu VztbKpmWjw6h0VkDDqTBOiVaWP29qg6arvApSZmKPwVX7r4Z5wm8MR0CI4Ibaepbk7JmSMJIYgnea BiKHFs1s6lcYlvMC7PplksdiAm/46rv7/lwE5V+dszHmcsdmdLtYNqirCe7R1X2o6MCFlr0bj0VNX G3+N+JdoLsFxUOpLA9+t+uHTKLsp7hOl3z5FLNSqG+XkXX33/C9pqnIvse9fHjugysF+ilMEOpuWA r1uXFRRw==; Received: from 2a02-8389-2341-5b80-3177-e4c1-2108-f294.cable.dynamic.v6.surfer.at ([2a02:8389:2341:5b80:3177:e4c1:2108:f294] helo=localhost) by bombadil.infradead.org with esmtpsa (Exim 4.97.1 #2 (Red Hat Linux)) id 1sBdKj-0000000Ft6i-2WfQ; Mon, 27 May 2024 16:36:26 +0000 From: Christoph Hellwig To: Trond Myklebust , Anna Schumaker , Matthew Wilcox Cc: linux-nfs@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org Subject: [PATCH 2/2] nfs: add support for large folios Date: Mon, 27 May 2024 18:36:09 +0200 Message-ID: <20240527163616.1135968-3-hch@lst.de> X-Mailer: git-send-email 2.43.0 In-Reply-To: <20240527163616.1135968-1-hch@lst.de> References: <20240527163616.1135968-1-hch@lst.de> Precedence: bulk X-Mailing-List: linux-nfs@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 X-SRS-Rewrite: SMTP reverse-path rewritten from by bombadil.infradead.org. See http://www.infradead.org/rpr.html NFS already is void of folio size assumption, so just pass the chunk size to __filemap_get_folio and set the large folio address_space flag for all regular files. Signed-off-by: Christoph Hellwig --- fs/nfs/file.c | 4 +++- fs/nfs/inode.c | 1 + 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/fs/nfs/file.c b/fs/nfs/file.c index 6bd127e6683dce..7f1295475a90fd 100644 --- a/fs/nfs/file.c +++ b/fs/nfs/file.c @@ -339,6 +339,7 @@ static int nfs_write_begin(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, struct page **pagep, void **fsdata) { + fgf_t fgp = FGP_WRITEBEGIN; struct folio *folio; int once_thru = 0; int ret; @@ -346,8 +347,9 @@ static int nfs_write_begin(struct file *file, struct address_space *mapping, dfprintk(PAGECACHE, "NFS: write_begin(%pD2(%lu), %u@%lld)\n", file, mapping->host->i_ino, len, (long long) pos); + fgp |= fgf_set_order(len); start: - folio = __filemap_get_folio(mapping, pos >> PAGE_SHIFT, FGP_WRITEBEGIN, + folio = __filemap_get_folio(mapping, pos >> PAGE_SHIFT, fgp, mapping_gfp_mask(mapping)); if (IS_ERR(folio)) return PTR_ERR(folio); diff --git a/fs/nfs/inode.c b/fs/nfs/inode.c index acef52ecb1bb7e..6d185af4cb29d4 100644 --- a/fs/nfs/inode.c +++ b/fs/nfs/inode.c @@ -491,6 +491,7 @@ nfs_fhget(struct super_block *sb, struct nfs_fh *fh, struct nfs_fattr *fattr) inode->i_fop = NFS_SB(sb)->nfs_client->rpc_ops->file_ops; inode->i_data.a_ops = &nfs_file_aops; nfs_inode_init_regular(nfsi); + mapping_set_large_folios(inode->i_mapping); } else if (S_ISDIR(inode->i_mode)) { inode->i_op = NFS_SB(sb)->nfs_client->rpc_ops->dir_inode_ops; inode->i_fop = &nfs_dir_operations;