@@ -768,15 +768,7 @@ struct cl_page {
enum cl_page_type cp_type:CP_TYPE_BITS;
unsigned int cp_defer_uptodate:1,
cp_ra_updated:1,
- cp_ra_used:1,
- /* fault page read grab extra referece */
- cp_fault_ref:1,
- /**
- * if fault page got delete before returned to
- * filemap_fault(), defer the vmpage detach/put
- * until filemap_fault() has been handled.
- */
- cp_defer_detach:1;
+ cp_ra_used:1;
/* which slab kmem index this memory allocated from */
short int cp_kmem_index;
@@ -2401,11 +2393,6 @@ int cl_io_lru_reserve(const struct lu_env *env, struct cl_io *io,
int cl_io_read_ahead(const struct lu_env *env, struct cl_io *io,
pgoff_t start, struct cl_read_ahead *ra);
-static inline int cl_io_is_pagefault(const struct cl_io *io)
-{
- return io->ci_type == CIT_FAULT && !io->u.ci_fault.ft_mkwrite;
-}
-
/**
* True, if @io is an O_APPEND write(2).
*/
@@ -1947,15 +1947,7 @@ int ll_readpage(struct file *file, struct page *vmpage)
unlock_page(vmpage);
result = 0;
}
- if (cl_io_is_pagefault(io) && result == 0) {
- /**
- * page fault, retain the cl_page reference until
- * vvp_io_kernel_fault() release it.
- */
- page->cp_fault_ref = 1;
- } else {
- cl_page_put(env, page);
- }
+ cl_page_put(env, page);
} else {
unlock_page(vmpage);
result = PTR_ERR(page);
@@ -1302,41 +1302,14 @@ static void vvp_io_rw_end(const struct lu_env *env,
trunc_sem_up_read(&lli->lli_trunc_sem);
}
-static void detach_and_deref_page(struct cl_page *clp, struct page *vmpage)
-{
- if (!clp->cp_defer_detach)
- return;
-
- /**
- * cl_page_delete0() took a vmpage reference, but not unlink the vmpage
- * from its cl_page.
- */
- clp->cp_defer_detach = 0;
- ClearPagePrivate(vmpage);
- vmpage->private = 0;
-
- put_page(vmpage);
- refcount_dec(&clp->cp_ref);
-}
-
-static int vvp_io_kernel_fault(const struct lu_env *env,
- struct vvp_fault_io *cfio)
+static int vvp_io_kernel_fault(struct vvp_fault_io *cfio)
{
struct vm_fault *vmf = cfio->ft_vmf;
- struct file *vmff = cfio->ft_vma->vm_file;
- struct address_space *mapping = vmff->f_mapping;
- struct inode *inode = mapping->host;
- struct page *vmpage = NULL;
- struct cl_page *clp = NULL;
- int rc = 0;
- ll_inode_size_lock(inode);
-retry:
cfio->ft_flags = filemap_fault(vmf);
cfio->ft_flags_valid = 1;
if (vmf->page) {
- /* success, vmpage is locked */
CDEBUG(D_PAGE,
"page %p map %p index %lu flags %lx count %u priv %0lx: got addr %p type NOPAGE\n",
vmf->page, vmf->page->mapping, vmf->page->index,
@@ -1348,105 +1321,24 @@ static int vvp_io_kernel_fault(const struct lu_env *env,
}
cfio->ft_vmpage = vmf->page;
-
- /**
- * ll_filemap_fault()->ll_readpage() could get an extra cl_page
- * reference. So we have to get the cl_page's to check its
- * cp_fault_ref and drop the reference later.
- */
- clp = cl_vmpage_page(vmf->page, NULL);
-
- goto unlock;
- }
-
- /* filemap_fault() fails, vmpage is not locked */
- if (!clp) {
- vmpage = find_get_page(mapping, vmf->pgoff);
- if (vmpage) {
- lock_page(vmpage);
- clp = cl_vmpage_page(vmpage, NULL);
- unlock_page(vmpage);
- }
+ return 0;
}
if (cfio->ft_flags & (VM_FAULT_SIGBUS | VM_FAULT_SIGSEGV)) {
- pgoff_t max_idx;
-
- /**
- * ll_filemap_fault()->ll_readpage() could fill vmpage
- * correctly, and unlock the vmpage, while memory pressure or
- * truncate could detach cl_page from vmpage, and kernel
- * filemap_fault() will wait_on_page_locked(vmpage) and find
- * out that the vmpage has been cleared its uptodate bit,
- * so it returns VM_FAULT_SIGBUS.
- *
- * In this case, we'd retry the filemap_fault()->ll_readpage()
- * to rebuild the cl_page and fill vmpage with uptodated data.
- */
- if (likely(vmpage)) {
- bool need_retry = false;
-
- if (clp) {
- if (clp->cp_defer_detach) {
- detach_and_deref_page(clp, vmpage);
- /**
- * check i_size to make sure it's not
- * over EOF, we don't want to call
- * filemap_fault() repeatedly since it
- * returns VM_FAULT_SIGBUS without even
- * trying if vmf->pgoff is over EOF.
- */
- max_idx = DIV_ROUND_UP(i_size_read(inode),
- PAGE_SIZE);
- if (vmf->pgoff < max_idx)
- need_retry = true;
- }
- if (clp->cp_fault_ref) {
- clp->cp_fault_ref = 0;
- /* ref not released in ll_readpage() */
- cl_page_put(env, clp);
- }
- if (need_retry)
- goto retry;
- }
- }
-
CDEBUG(D_PAGE, "got addr %p - SIGBUS\n", (void *)vmf->address);
- rc = -EFAULT;
- goto unlock;
+ return -EFAULT;
}
if (cfio->ft_flags & VM_FAULT_OOM) {
CDEBUG(D_PAGE, "got addr %p - OOM\n", (void *)vmf->address);
- rc = -ENOMEM;
- goto unlock;
+ return -ENOMEM;
}
- if (cfio->ft_flags & VM_FAULT_RETRY) {
- rc = -EAGAIN;
- goto unlock;
- }
+ if (cfio->ft_flags & VM_FAULT_RETRY)
+ return -EAGAIN;
CERROR("Unknown error in page fault %d!\n", cfio->ft_flags);
- rc = -EINVAL;
-unlock:
- ll_inode_size_unlock(inode);
- if (clp) {
- if (clp->cp_defer_detach && vmpage)
- detach_and_deref_page(clp, vmpage);
-
- /* additional cl_page ref has been taken in ll_readpage() */
- if (clp->cp_fault_ref) {
- clp->cp_fault_ref = 0;
- /* ref not released in ll_readpage() */
- cl_page_put(env, clp);
- }
- /* ref taken in this function */
- cl_page_put(env, clp);
- }
- if (vmpage)
- put_page(vmpage);
- return rc;
+ return -EINVAL;
}
static void mkwrite_commit_callback(const struct lu_env *env, struct cl_io *io,
@@ -1486,7 +1378,7 @@ static int vvp_io_fault_start(const struct lu_env *env,
LASSERT(cfio->ft_vmpage);
lock_page(cfio->ft_vmpage);
} else {
- result = vvp_io_kernel_fault(env, cfio);
+ result = vvp_io_kernel_fault(cfio);
if (result != 0)
return result;
}
@@ -104,11 +104,6 @@ static void vvp_page_completion_read(const struct lu_env *env,
ll_ra_count_put(ll_i2sbi(inode), 1);
if (ioret == 0) {
- /**
- * cp_defer_uptodate is used for readahead page, and the
- * vmpage Uptodate bit is deferred to set in ll_readpage/
- * ll_io_read_page.
- */
if (!cp->cp_defer_uptodate)
SetPageUptodate(vmpage);
} else if (cp->cp_defer_uptodate) {
@@ -725,35 +725,16 @@ static void __cl_page_delete(const struct lu_env *env, struct cl_page *cp)
LASSERT(PageLocked(vmpage));
LASSERT((struct cl_page *)vmpage->private == cp);
- /**
- * clear vmpage uptodate bit, since ll_read_ahead_pages()->
- * ll_read_ahead_page() could pick up this stale vmpage and
- * take it as uptodated.
- */
- ClearPageUptodate(vmpage);
- /**
- * vvp_io_kernel_fault()->ll_readpage() set cp_fault_ref
- * and need it to check cl_page to retry the page fault read.
+ /* Drop the reference count held in vvp_page_init */
+ refcount_dec(&cp->cp_ref);
+ ClearPagePrivate(vmpage);
+ vmpage->private = 0;
+
+ /*
+ * The reference from vmpage to cl_page is removed,
+ * but the reference back is still here. It is removed
+ * later in cl_page_free().
*/
- if (cp->cp_fault_ref) {
- cp->cp_defer_detach = 1;
- /**
- * get a vmpage reference, so that filemap_fault()
- * won't free it from pagecache.
- */
- get_page(vmpage);
- } else {
- /* Drop the reference count held in vvp_page_init */
- refcount_dec(&cp->cp_ref);
- ClearPagePrivate(vmpage);
- vmpage->private = 0;
-
- /*
- * The reference from vmpage to cl_page is removed,
- * but the reference back is still here. It is removed
- * later in cl_page_free().
- */
- }
}
}