From patchwork Fri Oct 8 14:06:22 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Charan Teja Kalla X-Patchwork-Id: 12545497 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 5FE29C433F5 for ; Fri, 8 Oct 2021 14:07:22 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 0712960F58 for ; Fri, 8 Oct 2021 14:07:22 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.4.1 mail.kernel.org 0712960F58 Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=codeaurora.org Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=kvack.org Received: by kanga.kvack.org (Postfix) id 92F91900002; Fri, 8 Oct 2021 10:07:21 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 8DFA86B0072; Fri, 8 Oct 2021 10:07:21 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 7CE09900002; Fri, 8 Oct 2021 10:07:21 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0102.hostedemail.com [216.40.44.102]) by kanga.kvack.org (Postfix) with ESMTP id 6DAC66B0071 for ; Fri, 8 Oct 2021 10:07:21 -0400 (EDT) Received: from smtpin12.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay04.hostedemail.com (Postfix) with ESMTP id 1790439B0C for ; Fri, 8 Oct 2021 14:07:21 +0000 (UTC) X-FDA: 78673447482.12.DD74B7A Received: from so254-9.mailgun.net (so254-9.mailgun.net [198.61.254.9]) by imf01.hostedemail.com (Postfix) with ESMTP id 9EE315078955 for ; Fri, 8 Oct 2021 14:07:18 +0000 (UTC) DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mg.codeaurora.org; q=dns/txt; s=smtp; t=1633702040; h=Message-Id: Date: Subject: Cc: To: From: Sender; bh=z5HFXO447DX2xV076VYmLHupBfI2lHWsPlxwcYyE54k=; b=Y29r+fs++dgnrKErg19jYLkL6XNf+7Si0qY2jbk03xUR0XuddmhFK47OyYyt6rOaXxoJozkY deSnwXswoYW5JVYu8eRSSfI6m8RYwvQiXpavQeNBjKPHd0JRTbJS8gW2CiIOH0Ls2XtdZFEM 5HtnOt1PoNLOD2QLoM5zWHB1UMM= X-Mailgun-Sending-Ip: 198.61.254.9 X-Mailgun-Sid: WyIwY2Q3OCIsICJsaW51eC1tbUBrdmFjay5vcmciLCAiYmU5ZTRhIl0= Received: from smtp.codeaurora.org (ec2-35-166-182-171.us-west-2.compute.amazonaws.com [35.166.182.171]) by smtp-out-n07.prod.us-west-2.postgun.com with SMTP id 61605092ff0285fb0a44b9d1 (version=TLS1.2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256); Fri, 08 Oct 2021 14:07:14 GMT Received: by smtp.codeaurora.org (Postfix, from userid 1001) id 239A6C43616; Fri, 8 Oct 2021 14:07:14 +0000 (UTC) Received: from hu-charante-hyd.qualcomm.com (unknown [202.46.22.19]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-SHA256 (128/128 bits)) (No client certificate requested) (Authenticated sender: charante) by smtp.codeaurora.org (Postfix) with ESMTPSA id 3F340C43460; Fri, 8 Oct 2021 14:07:09 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.4.1 smtp.codeaurora.org 3F340C43460 From: Charan Teja Reddy To: hughd@google.com, akpm@linux-foundation.org, vbabka@suse.cz, rientjes@google.com, david@redhat.com, mhocko@suse.com, linux-mm@kvack.org Cc: linux-kernel@vger.kernel.org, Charan Teja Reddy Subject: [PATCH] mm: shmem: implement POSIX_FADV_[WILL|DONT]NEED for shmem Date: Fri, 8 Oct 2021 19:36:22 +0530 Message-Id: <1633701982-22302-1-git-send-email-charante@codeaurora.org> X-Mailer: git-send-email 2.7.4 Authentication-Results: imf01.hostedemail.com; dkim=pass header.d=mg.codeaurora.org header.s=smtp header.b=Y29r+fs+; spf=pass (imf01.hostedemail.com: domain of "bounce+d06763.be9e4a-linux-mm=kvack.org@mg.codeaurora.org" designates 198.61.254.9 as permitted sender) smtp.mailfrom="bounce+d06763.be9e4a-linux-mm=kvack.org@mg.codeaurora.org"; dmarc=none X-Rspamd-Server: rspam04 X-Rspamd-Queue-Id: 9EE315078955 X-Stat-Signature: 7ks6ttjjt3fp9tks8zeowoci6dni3to9 X-HE-Tag: 1633702038-290646 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: Currently fadvise(2) is supported only for the files that doesn't associated with noop_backing_dev_info thus for the files, like shmem, fadvise results into NOP. But then there is file_operations->fadvise() that lets the file systems to implement their own fadvise implementation. Use this support to implement some of the POSIX_FADV_XXX functionality for shmem files. This patch aims to implement POSIX_FADV_WILLNEED and POSIX_FADV_DONTNEED advices to shmem files which can be helpful for the drivers who may want to manage the shmem pages of the files that are created through shmem_file_setup[_with_mnt](). An example usecase may be like, driver can create the shmem file of the size equal to its requirements and map the pages for DMA and then pass the fd to user. The user who knows well about the usage of these pages can now decide when these pages are not required push them to swap through DONTNEED thus free up memory well in advance rather than relying on the reclaim and use WILLNEED when it decide that they are useful in the near future. IOW, it lets the clients to free up/read the memory when it wants to. Another usecase is that GEM objets which are currenlty allocated and managed through shmem files can use vfs_fadvise(DONT|WILLNEED) on shmem fd when the driver comes to know(like through some hints from user space) that GEM objects are not going to use/will need in the near future. Signed-off-by: Charan Teja Reddy --- mm/shmem.c | 139 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 139 insertions(+) diff --git a/mm/shmem.c b/mm/shmem.c index 70d9ce2..ab7ea33 100644 --- a/mm/shmem.c +++ b/mm/shmem.c @@ -38,6 +38,8 @@ #include #include #include +#include +#include #include /* for arch/microblaze update_mmu_cache() */ @@ -2792,6 +2794,142 @@ static long shmem_fallocate(struct file *file, int mode, loff_t offset, return error; } +static int shmem_fadvise_dontneed(struct address_space *mapping, loff_t start, + loff_t end) +{ + int ret; + struct writeback_control wbc = { + .sync_mode = WB_SYNC_NONE, + .nr_to_write = LONG_MAX, + .range_start = 0, + .range_end = LLONG_MAX, + .for_reclaim = 1, + }; + struct page *page; + + XA_STATE(xas, &mapping->i_pages, start); + if (!shmem_mapping(mapping)) + return -EINVAL; + + if (!total_swap_pages) + return 0; + + lru_add_drain(); + + rcu_read_lock(); + xas_for_each(&xas, page, end) { + if (xas_retry(&xas, page)) + continue; + if (xa_is_value(page)) + continue; + if (isolate_lru_page(page)) + continue; + + inc_node_page_state(page, NR_ISOLATED_ANON + + page_is_file_lru(page)); + lock_page(page); + ClearPageDirty(page); + SetPageReclaim(page); + ret = shmem_writepage(page, &wbc); + if (!PageWriteback(page)) + ClearPageReclaim(page); + if (ret) { + unlock_page(page); + putback_lru_page(page); + dec_node_page_state(page, NR_ISOLATED_ANON + + page_is_file_lru(page)); + continue; + } + + /* + * shmem_writepage() place the page in the swapcache. + * Delete the page from the swapcache and release the + * page. + */ + lock_page(page); + delete_from_swap_cache(page); + unlock_page(page); + dec_node_page_state(page, NR_ISOLATED_ANON + + page_is_file_lru(page)); + put_page(page); + if (need_resched()) { + xas_pause(&xas); + cond_resched_rcu(); + } + } + rcu_read_unlock(); + + return 0; +} + +static int shmem_fadvise_willneed(struct address_space *mapping, + pgoff_t start, pgoff_t long end) +{ + struct page *page; + + XA_STATE(xas, &mapping->i_pages, start); + rcu_read_lock(); + xas_for_each(&xas, page, end) { + if (!xa_is_value(page)) + continue; + page = shmem_read_mapping_page(mapping, xas.xa_index); + if (!IS_ERR(page)) + put_page(page); + + if (need_resched()) { + xas_pause(&xas); + cond_resched_rcu(); + } + } + rcu_read_unlock(); + + return 0; +} + +static int shmem_fadvise(struct file *file, loff_t offset, loff_t len, int advice) +{ + loff_t endbyte; + pgoff_t start_index; + pgoff_t end_index; + struct address_space *mapping; + int ret = 0; + + mapping = file->f_mapping; + if (!mapping || len < 0) + return -EINVAL; + + endbyte = (u64)offset + (u64)len; + if (!len || endbyte < len) + endbyte = -1; + else + endbyte--; + + + start_index = offset >> PAGE_SHIFT; + end_index = endbyte >> PAGE_SHIFT; + switch (advice) { + case POSIX_FADV_DONTNEED: + ret = shmem_fadvise_dontneed(mapping, start_index, end_index); + break; + case POSIX_FADV_WILLNEED: + ret = shmem_fadvise_willneed(mapping, start_index, end_index); + break; + case POSIX_FADV_NORMAL: + case POSIX_FADV_RANDOM: + case POSIX_FADV_SEQUENTIAL: + case POSIX_FADV_NOREUSE: + /* + * No bad return value, but ignore advice. May have to + * implement in future. + */ + break; + default: + return -EINVAL; + } + + return ret; +} + static int shmem_statfs(struct dentry *dentry, struct kstatfs *buf) { struct shmem_sb_info *sbinfo = SHMEM_SB(dentry->d_sb); @@ -3799,6 +3937,7 @@ static const struct file_operations shmem_file_operations = { .splice_write = iter_file_splice_write, .fallocate = shmem_fallocate, #endif + .fadvise = shmem_fadvise, }; static const struct inode_operations shmem_inode_operations = {