From patchwork Tue May 10 12:27:15 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Jan Kara X-Patchwork-Id: 9057181 Return-Path: X-Original-To: patchwork-linux-fsdevel@patchwork.kernel.org Delivered-To: patchwork-parsemail@patchwork1.web.kernel.org Received: from mail.kernel.org (mail.kernel.org [198.145.29.136]) by patchwork1.web.kernel.org (Postfix) with ESMTP id 0737A9F30C for ; Tue, 10 May 2016 12:27:29 +0000 (UTC) Received: from mail.kernel.org (localhost [127.0.0.1]) by mail.kernel.org (Postfix) with ESMTP id EED82200E8 for ; Tue, 10 May 2016 12:27:26 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id B738220145 for ; Tue, 10 May 2016 12:27:25 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751540AbcEJM1W (ORCPT ); Tue, 10 May 2016 08:27:22 -0400 Received: from mx2.suse.de ([195.135.220.15]:42830 "EHLO mx2.suse.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751430AbcEJM1V (ORCPT ); Tue, 10 May 2016 08:27:21 -0400 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (charybdis-ext.suse.de [195.135.220.254]) by mx2.suse.de (Postfix) with ESMTP id 70A15AAEF; Tue, 10 May 2016 12:27:18 +0000 (UTC) Received: by quack2.suse.cz (Postfix, from userid 1000) id CDE811E0759; Tue, 10 May 2016 14:27:15 +0200 (CEST) Date: Tue, 10 May 2016 14:27:15 +0200 From: Jan Kara To: Ross Zwisler Cc: Jan Kara , linux-fsdevel@vger.kernel.org, linux-ext4@vger.kernel.org, linux-mm@kvack.org, Dan Williams , linux-nvdimm@lists.01.org, Matthew Wilcox Subject: Re: [PATCH 16/18] dax: New fault locking Message-ID: <20160510122715.GK11897@quack2.suse.cz> References: <1461015341-20153-1-git-send-email-jack@suse.cz> <1461015341-20153-17-git-send-email-jack@suse.cz> <20160506041350.GA29628@linux.intel.com> MIME-Version: 1.0 Content-Disposition: inline In-Reply-To: <20160506041350.GA29628@linux.intel.com> User-Agent: Mutt/1.5.24 (2015-08-30) Sender: linux-fsdevel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-fsdevel@vger.kernel.org X-Spam-Status: No, score=-9.0 required=5.0 tests=BAYES_00, RCVD_IN_DNSWL_HI, RP_MATCHES_RCVD, T_TVD_MIME_EPI, UNPARSEABLE_RELAY autolearn=unavailable version=3.3.1 X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on mail.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP On Thu 05-05-16 22:13:50, Ross Zwisler wrote: > On Mon, Apr 18, 2016 at 11:35:39PM +0200, Jan Kara wrote: > > /* > > + * DAX radix tree locking > > + */ > > +struct exceptional_entry_key { > > + struct radix_tree_root *root; > > + unsigned long index; > > +}; > > I believe that we basically just need the struct exceptional_entry_key to > uniquely identify an entry, correct? I agree that we get this with the pair > [struct radix_tree_root, index], but we also get it with > [struct address_space, index], and we might want to use the latter here since > that's the pair that is used to look up the wait queue in > dax_entry_waitqueue(). Functionally I don't think it matters (correct me if > I'm wrong), but it makes for a nicer symmetry. OK, makes sense. Changed. > > +/* > > + * Find radix tree entry at given index. If it points to a page, return with > > + * the page locked. If it points to the exceptional entry, return with the > > + * radix tree entry locked. If the radix tree doesn't contain given index, > > + * create empty exceptional entry for the index and return with it locked. > > + * > > + * Note: Unlike filemap_fault() we don't honor FAULT_FLAG_RETRY flags. For > > + * persistent memory the benefit is doubtful. We can add that later if we can > > + * show it helps. > > + */ > > +static void *grab_mapping_entry(struct address_space *mapping, pgoff_t index) > > +{ > > + void *ret, **slot; > > + > > +restart: > > + spin_lock_irq(&mapping->tree_lock); > > + ret = get_unlocked_mapping_entry(mapping, index, &slot); > > + /* No entry for given index? Make sure radix tree is big enough. */ > > + if (!ret) { > > + int err; > > + > > + spin_unlock_irq(&mapping->tree_lock); > > + err = radix_tree_preload( > > + mapping_gfp_mask(mapping) & ~__GFP_HIGHMEM); > > In the conversation about v2 of this series you said: > > > Note that we take the hit for dropping the lock only if we really need to > > allocate new radix tree node so about once per 64 new entries. So it is not > > too bad. > > I think this is incorrect. We get here whenever we get a NULL return from > __radix_tree_lookup(). I believe that this happens if we don't have a node, > in which case we need an allocation, but I think it also happens in the case > where we do have a node and we just have a NULL slot in that node. > > For the behavior you're looking for (only preload if you need to do an > allocation), you probably need to check the 'slot' we get back from > get_unlocked_mapping_entry(), yea? You are correct. However currently __radix_tree_lookup() doesn't return a slot pointer if entry was not found so it is not easy to fix. So I'd leave the code as is for now and we can later optimize the case where we don't need to grow the radix tree... > > > +/* > > + * Delete exceptional DAX entry at @index from @mapping. Wait for radix tree > > + * entry to get unlocked before deleting it. > > + */ > > +int dax_delete_mapping_entry(struct address_space *mapping, pgoff_t index) > > +{ > > + void *entry; > > + > > + spin_lock_irq(&mapping->tree_lock); > > + entry = get_unlocked_mapping_entry(mapping, index, NULL); > > + /* > > + * Caller should make sure radix tree modifications don't race and > > + * we have seen exceptional entry here before. > > + */ > > + if (WARN_ON_ONCE(!entry || !radix_tree_exceptional_entry(entry))) { > > dax_delete_mapping_entry() is only called from clear_exceptional_entry(). > With this new code we've changed the behavior of that call path a little. > > In the various places where clear_exceptional_entry() is called, the code > batches up a bunch of entries in a pvec via pagevec_lookup_entries(). We > don't hold the mapping->tree_lock between the time this lookup happens and the > time that the entry is passed to clear_exceptional_entry(). This is why the > old code did a verification that the entry passed in matched what was still > currently present in the radix tree. This was done in the DAX case via > radix_tree_delete_item(), and it was open coded in clear_exceptional_entry() > for the page cache case. In both cases if the entry didn't match what was > currently in the tree, we bailed without doing anything. > > This new code doesn't verify against the 'entry' passed to > clear_exceptional_entry(), but instead makes sure it is an exceptional entry > before removing, and if not it does a WARN_ON_ONCE(). > > This changes things because: > > a) If the exceptional entry changed, say from a plain lock entry to an actual > DAX entry, we wouldn't notice, and we would just clear the latter out. My > guess is that this is fine, I just wanted to call it out. > > b) If we have a non-exceptional entry here now, say because our lock entry has > been swapped out for a zero page, we will WARN_ON_ONCE() and return without a > removal. I think we may want to silence the WARN_ON_ONCE(), as I believe this > could happen during normal operation and we don't want to scare anyone. :) So your concerns are exactly why I have added a comment to dax_delete_mapping_entry() that: /* * Caller should make sure radix tree modifications don't race and * we have seen exceptional entry here before. */ The thing is dax_delete_mapping_entry() is called only from truncate / punch hole path. Those should hold i_mmap_sem for writing and thus there should be no modifications of the radix tree. If anything changes, between what truncate_inode_pages() (or similar functions) finds and what dax_delete_mapping_entry() sees, we have a locking bug and I want to know about it :). Any suggestion how I should expand the comment so that this is clearer? > > +/* > > * The user has performed a load from a hole in the file. Allocating > > * a new page in the file would cause excessive storage usage for > > * workloads with sparse files. We allocate a page cache page instead. > > @@ -307,15 +584,24 @@ EXPORT_SYMBOL_GPL(dax_do_io); > > * otherwise it will simply fall out of the page cache under memory > > * pressure without ever having been dirtied. > > */ > > -static int dax_load_hole(struct address_space *mapping, struct page *page, > > - struct vm_fault *vmf) > > +static int dax_load_hole(struct address_space *mapping, void *entry, > > + struct vm_fault *vmf) > > { > > - if (!page) > > - page = find_or_create_page(mapping, vmf->pgoff, > > - GFP_KERNEL | __GFP_ZERO); > > - if (!page) > > - return VM_FAULT_OOM; > > + struct page *page; > > + > > + /* Hole page already exists? Return it... */ > > + if (!radix_tree_exceptional_entry(entry)) { > > + vmf->page = entry; > > + return VM_FAULT_LOCKED; > > + } > > > > + /* This will replace locked radix tree entry with a hole page */ > > + page = find_or_create_page(mapping, vmf->pgoff, > > + vmf->gfp_mask | __GFP_ZERO); > > This replacement happens via page_cache_tree_insert(), correct? In this case, > who wakes up anyone waiting on the old lock entry that we just killed? In the > non-hole case we would traverse through put_locked_mapping_entry(), but I > don't see that in the hole case. Ha, good catch. We miss the wakeup. Fixed. Attached is the diff resulting from your review of this patch. I still have to hunt down that strange interaction with workingset code you've reported... Honza diff --git a/fs/dax.c b/fs/dax.c index 26798cdc6789..2913a82dd68d 100644 --- a/fs/dax.c +++ b/fs/dax.c @@ -327,7 +327,7 @@ EXPORT_SYMBOL_GPL(dax_do_io); * DAX radix tree locking */ struct exceptional_entry_key { - struct radix_tree_root *root; + struct address_space *mapping; unsigned long index; }; @@ -343,7 +343,8 @@ static int wake_exceptional_entry_func(wait_queue_t *wait, unsigned int mode, struct wait_exceptional_entry_queue *ewait = container_of(wait, struct wait_exceptional_entry_queue, wait); - if (key->root != ewait->key.root || key->index != ewait->key.index) + if (key->mapping != ewait->key.mapping || + key->index != ewait->key.index) return 0; return autoremove_wake_function(wait, mode, sync, NULL); } @@ -489,8 +490,8 @@ restart: return ret; } -static void wake_mapping_entry_waiter(struct address_space *mapping, - pgoff_t index, bool wake_all) +void dax_wake_mapping_entry_waiter(struct address_space *mapping, + pgoff_t index, bool wake_all) { wait_queue_head_t *wq = dax_entry_waitqueue(mapping, index); @@ -503,7 +504,7 @@ static void wake_mapping_entry_waiter(struct address_space *mapping, if (waitqueue_active(wq)) { struct exceptional_entry_key key; - key.root = &mapping->page_tree; + key.mapping = mapping; key.index = index; __wake_up(wq, TASK_NORMAL, wake_all ? 0 : 1, &key); } @@ -522,7 +523,7 @@ static void unlock_mapping_entry(struct address_space *mapping, pgoff_t index) } unlock_slot(mapping, slot); spin_unlock_irq(&mapping->tree_lock); - wake_mapping_entry_waiter(mapping, index, false); + dax_wake_mapping_entry_waiter(mapping, index, false); } static void put_locked_mapping_entry(struct address_space *mapping, @@ -547,7 +548,7 @@ static void put_unlocked_mapping_entry(struct address_space *mapping, return; /* We have to wake up next waiter for the radix tree entry lock */ - wake_mapping_entry_waiter(mapping, index, false); + dax_wake_mapping_entry_waiter(mapping, index, false); } /* @@ -571,7 +572,7 @@ int dax_delete_mapping_entry(struct address_space *mapping, pgoff_t index) radix_tree_delete(&mapping->page_tree, index); mapping->nrexceptional--; spin_unlock_irq(&mapping->tree_lock); - wake_mapping_entry_waiter(mapping, index, true); + dax_wake_mapping_entry_waiter(mapping, index, true); return 1; } diff --git a/include/linux/dax.h b/include/linux/dax.h index be40ec13d469..d3d788b44d66 100644 --- a/include/linux/dax.h +++ b/include/linux/dax.h @@ -17,6 +17,8 @@ int dax_truncate_page(struct inode *, loff_t from, get_block_t); int dax_fault(struct vm_area_struct *, struct vm_fault *, get_block_t); int __dax_fault(struct vm_area_struct *, struct vm_fault *, get_block_t); int dax_delete_mapping_entry(struct address_space *mapping, pgoff_t index); +void dax_wake_mapping_entry_waiter(struct address_space *mapping, + pgoff_t index, bool wake_all); #ifdef CONFIG_FS_DAX struct page *read_dax_sector(struct block_device *bdev, sector_t n); diff --git a/mm/filemap.c b/mm/filemap.c index 3effd5c8f2f6..6d42525a68eb 100644 --- a/mm/filemap.c +++ b/mm/filemap.c @@ -608,6 +608,9 @@ static int page_cache_tree_insert(struct address_space *mapping, WARN_ON_ONCE(p != (void *)(RADIX_TREE_EXCEPTIONAL_ENTRY | RADIX_DAX_ENTRY_LOCK)); + /* Wakeup waiters for exceptional entry lock */ + dax_wake_mapping_entry_waiter(mapping, page->index, + false); } } radix_tree_replace_slot(slot, page);