Message ID | 1445557283.17208.30.camel@intel.com (mailing list archive) |
---|---|
State | Superseded |
Headers | show |
On Thu 22-10-15 23:41:27, Williams, Dan J wrote: > On Thu, 2015-10-22 at 23:08 +0200, Jan Kara wrote: > > On Thu 22-10-15 16:05:46, Williams, Dan J wrote: > > > On Thu, 2015-10-22 at 11:35 +0200, Jan Kara wrote: > > > > On Thu 22-10-15 02:42:11, Dan Williams wrote: > > > > > If an application wants exclusive access to all of the persistent memory > > > > > provided by an NVDIMM namespace it can use this raw-block-dax facility > > > > > to forgo establishing a filesystem. This capability is targeted > > > > > primarily to hypervisors wanting to provision persistent memory for > > > > > guests. > > > > > > > > > > Cc: Jan Kara <jack@suse.cz> > > > > > Cc: Jeff Moyer <jmoyer@redhat.com> > > > > > Cc: Christoph Hellwig <hch@lst.de> > > > > > Cc: Dave Chinner <david@fromorbit.com> > > > > > Cc: Andrew Morton <akpm@linux-foundation.org> > > > > > Cc: Ross Zwisler <ross.zwisler@linux.intel.com> > > > > > Signed-off-by: Dan Williams <dan.j.williams@intel.com> > > > > > --- > > > > > fs/block_dev.c | 54 +++++++++++++++++++++++++++++++++++++++++++++++++++++- > > > > > 1 file changed, 53 insertions(+), 1 deletion(-) > > > > > > > > > > diff --git a/fs/block_dev.c b/fs/block_dev.c > > > > > index 3255dcec96b4..c27cd1a21a13 100644 > > > > > --- a/fs/block_dev.c > > > > > +++ b/fs/block_dev.c > > > > > @@ -1687,13 +1687,65 @@ static const struct address_space_operations def_blk_aops = { > > > > > .is_dirty_writeback = buffer_check_dirty_writeback, > > > > > }; > > > > > > > > > > +#ifdef CONFIG_FS_DAX > > > > > +/* > > > > > + * In the raw block case we do not need to contend with truncation nor > > > > > + * unwritten file extents. Without those concerns there is no need for > > > > > + * additional locking beyond the mmap_sem context that these routines > > > > > + * are already executing under. > > > > > + * > > > > > + * Note, there is no protection if the block device is dynamically > > > > > + * resized (partition grow/shrink) during a fault. A stable block device > > > > > + * size is already not enforced in the blkdev_direct_IO path. > > > > > + * > > > > > + * For DAX, it is the responsibility of the block device driver to > > > > > + * ensure the whole-disk device size is stable while requests are in > > > > > + * flight. > > > > > + * > > > > > + * Finally, these paths do not synchronize against freezing > > > > > + * (sb_start_pagefault(), etc...) since bdev_sops does not support > > > > > + * freezing. > > > > > > > > Well, for devices freezing is handled directly in the block layer code > > > > (blk_stop_queue()) since there's no need to put some metadata structures > > > > into a consistent state. So the comment about bdev_sops is somewhat > > > > strange. > > > > > > This text was aimed at the request from Ross to document the differences > > > vs the generic_file_mmap() path. Is the following incremental change > > > more clear? > > > > Well, not really. I thought you'd just delete that paragraph :) The thing > > is: When doing IO directly to the block device, it makes no sense to look > > at a filesystem on top of it - hopefully there is none since you'd be > > corrupting it. So the paragraph that would make sense to me would be: > > > > * Finally, in contrast to filemap_page_mkwrite(), we don't bother calling > > * sb_start_pagefault(). There is no filesystem which could be frozen here > > * and when bdev gets frozen, IO gets blocked in the request queue. > > > > But when spelled out like this, I've realized that with DAX, this blocking > > of requests in the request queue doesn't really block the IO to the device. > > So block device freezing (aka blk_queue_stop()) doesn't work reliably with > > DAX. That should be fixed but it's not easy as the only way to do that > > would be to hook into blk_stop_queue() and unmap (or at least > > write-protect) all the mappings of the device. Ugh... > > > > Ugh2: Now I realized that DAX mmap isn't safe wrt fs freezing even for > > filesystems since there's nothing which writeprotects pages that are > > writeably mapped. In normal path, page writeback does this but that doesn't > > happen for DAX. I remember we once talked about this but it got lost. > > We need something like walk all filesystem inodes during fs freeze and > > writeprotect all pages that are mapped. But that's going to be slow... > > This sounds suspiciously like what I'm planning to do for the device > teardown path when we've dynamically allocated struct page. The backing > memory for those pages is freed when the driver runs its ->remove() > path, so we have to be sure there are no outstanding references to them. > > My current proposal for the teardown case, that we might re-purpose for > this freeze case, is below. It relies on the percpu_ref in the > request_queue to block new faults and then uses truncate_pagecache() to > teardown mappings. However, this assumes we've inserted pages into the > address_space radix at fault, which we don't currently do... Well, for the freeze case it is enough to call unmap_mapping_range() for each inode->i_mapping on the frozen filesystem. Struct page or presence in radix tree isn't needed for that to work. Less intrusive solution would be to do what unmap_mapping_range() does but writeprotect all the ptes instead of invalidating them. But that would require some more coding. > In general, as this page-backed-pmem support lands upstream, I'm of the > opinion that the page-less DAX support be deprecated/disabled > unless/until it can be made as functionally capable as the page-enabled > paths. I didn't get to reading those patches yet so I may be behind on what has been agreed on. So far it seemed to me that we can get most of the functionality work without struct page so that would be preferable so that we don't have to allocate those pages, no? For stuff like get_user_pages() allocating struct page is probably the least painful path so I agree with struct page there. But that is relatively rare... We can talk about this at KS. Honza
diff --git a/drivers/nvdimm/pmem.c b/drivers/nvdimm/pmem.c index f7acce594fa0..2c9aebbc3fea 100644 --- a/drivers/nvdimm/pmem.c +++ b/drivers/nvdimm/pmem.c @@ -24,12 +24,15 @@ #include <linux/memory_hotplug.h> #include <linux/moduleparam.h> #include <linux/vmalloc.h> +#include <linux/async.h> #include <linux/slab.h> #include <linux/pmem.h> #include <linux/nd.h> #include "pfn.h" #include "nd.h" +static ASYNC_DOMAIN_EXCLUSIVE(async_pmem); + struct pmem_device { struct request_queue *pmem_queue; struct gendisk *pmem_disk; @@ -164,14 +167,43 @@ static struct pmem_device *pmem_alloc(struct device *dev, return pmem; } -static void pmem_detach_disk(struct pmem_device *pmem) + +static void async_blk_cleanup_queue(void *data, async_cookie_t cookie) +{ + struct pmem_device *pmem = data; + + blk_cleanup_queue(pmem->pmem_queue); +} + +static void pmem_detach_disk(struct device *dev) { + struct pmem_device *pmem = dev_get_drvdata(dev); + struct request_queue *q = pmem->pmem_queue; + if (!pmem->pmem_disk) return; del_gendisk(pmem->pmem_disk); put_disk(pmem->pmem_disk); - blk_cleanup_queue(pmem->pmem_queue); + async_schedule_domain(async_blk_cleanup_queue, pmem, &async_pmem); + + if (pmem->pfn_flags & PFN_MAP) { + /* + * Wait for queue to go dead so that we know no new + * references will be taken against the pages allocated + * by devm_memremap_pages(). + */ + blk_wait_queue_dead(q); + + /* + * Manually release the page mapping so that + * blk_cleanup_queue() can complete queue draining. + */ + devm_memunmap_pages(dev, (void __force *) pmem->virt_addr); + } + + /* Wait for blk_cleanup_queue() to finish */ + async_synchronize_full_domain(&async_pmem); } static int pmem_attach_disk(struct device *dev, @@ -299,11 +331,9 @@ static int nd_pfn_init(struct nd_pfn *nd_pfn) static int nvdimm_namespace_detach_pfn(struct nd_namespace_common *ndns) { struct nd_pfn *nd_pfn = to_nd_pfn(ndns->claim); - struct pmem_device *pmem; /* free pmem disk */ - pmem = dev_get_drvdata(&nd_pfn->dev); - pmem_detach_disk(pmem); + pmem_detach_disk(&nd_pfn->dev); /* release nd_pfn resources */ kfree(nd_pfn->pfn_sb); @@ -446,7 +476,7 @@ static int nd_pmem_remove(struct device *dev) else if (is_nd_pfn(dev)) nvdimm_namespace_detach_pfn(pmem->ndns); else - pmem_detach_disk(pmem); + pmem_detach_disk(dev); return 0; } diff --git a/fs/dax.c b/fs/dax.c index 8d756562fcf0..0bc9b315d16f 100644 --- a/fs/dax.c +++ b/fs/dax.c @@ -46,6 +46,7 @@ static void __pmem *__dax_map_atomic(struct block_device *bdev, sector_t sector, blk_queue_exit(q); return (void __pmem *) ERR_PTR(rc); } + rcu_read_lock(); return addr; } @@ -62,6 +63,7 @@ static void dax_unmap_atomic(struct block_device *bdev, void __pmem *addr) if (IS_ERR(addr)) return; blk_queue_exit(bdev->bd_queue); + rcu_read_unlock(); } int dax_clear_blocks(struct inode *inode, sector_t block, long size) diff --git a/include/linux/mm.h b/include/linux/mm.h index a5b5267eae5b..294518ddf5bc 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -801,6 +801,7 @@ struct dev_pagemap { #ifdef CONFIG_ZONE_DEVICE struct dev_pagemap *__get_dev_pagemap(resource_size_t phys); +void devm_memunmap_pages(struct device *dev, void *addr); void *devm_memremap_pages(struct device *dev, struct resource *res, struct percpu_ref *ref, struct vmem_altmap *altmap); #else @@ -809,6 +810,10 @@ static inline struct dev_pagemap *__get_dev_pagemap(resource_size_t phys) return NULL; } +static inline void devm_memunmap_pages(struct device *dev, void *addr) +{ +} + static inline void *devm_memremap_pages(struct device *dev, struct resource *res, struct percpu_ref *ref, struct vmem_altmap *altmap) { diff --git a/kernel/memremap.c b/kernel/memremap.c index 4698071a1c43..ac74336e6d73 100644 --- a/kernel/memremap.c +++ b/kernel/memremap.c @@ -13,6 +13,7 @@ #include <linux/rculist.h> #include <linux/device.h> #include <linux/types.h> +#include <linux/fs.h> #include <linux/io.h> #include <linux/mm.h> #include <linux/memory_hotplug.h> @@ -187,10 +188,39 @@ static unsigned long pfn_end(struct dev_pagemap *pgmap) static void devm_memremap_pages_release(struct device *dev, void *data) { + unsigned long pfn; struct page_map *page_map = data; struct resource *res = &page_map->res; + struct address_space *mapping_prev = NULL; struct dev_pagemap *pgmap = &page_map->pgmap; + if (percpu_ref_tryget_live(pgmap->ref)) { + dev_WARN(dev, "%s: page mapping is still live!\n", __func__); + percpu_ref_put(pgmap->ref); + } + + /* flush in-flight dax_map_atomic() operations */ + synchronize_rcu(); + + for_each_device_pfn(pfn, pgmap) { + struct page *page = pfn_to_page(pfn); + struct address_space *mapping = page->mapping; + struct inode *inode = mapping ? mapping->host : NULL; + + dev_WARN_ONCE(dev, atomic_read(&page->_count) < 1, + "%s: ZONE_DEVICE page was freed!\n", __func__); + + if (!mapping || !inode || mapping == mapping_prev) { + dev_WARN_ONCE(dev, atomic_read(&page->_count) > 1, + "%s: unexpected elevated page count pfn: %lx\n", + __func__, pfn); + continue; + } + + truncate_pagecache(inode, 0); + mapping_prev = mapping; + } + /* pages are dead and unused, undo the arch mapping */ arch_remove_memory(res->start, resource_size(res)); dev_WARN_ONCE(dev, pgmap->altmap && pgmap->altmap->alloc, @@ -292,6 +322,24 @@ void *devm_memremap_pages(struct device *dev, struct resource *res, return __va(res->start); } EXPORT_SYMBOL(devm_memremap_pages); + +static int page_map_match(struct device *dev, void *res, void *match_data) +{ + struct page_map *page_map = res; + resource_size_t phys = *(resource_size_t *) match_data; + + return page_map->res.start == phys; +} + +void devm_memunmap_pages(struct device *dev, void *addr) +{ + resource_size_t start = __pa(addr); + + if (devres_release(dev, devm_memremap_pages_release, page_map_match, + &start) != 0) + dev_WARN(dev, "failed to find page map to release\n"); +} +EXPORT_SYMBOL(devm_memunmap_pages); #endif /* CONFIG_ZONE_DEVICE */ #ifdef CONFIG_SPARSEMEM_VMEMMAP