vma->vm_file->f_dentry->d_inode->i_ino, \
vma->vm_file->f_dentry->d_iname, ## arg); \
-struct page *ll_nopage(struct vm_area_struct *vma, unsigned long address,
- int *type);
-
static struct vm_operations_struct ll_file_vm_ops;
void policy_from_vma(ldlm_policy_data_t *policy,
struct lu_env *env;
struct cl_env_nest nest;
struct cl_io *io;
- struct page *page = NOPAGE_SIGBUS;
- struct vvp_io *vio = NULL;
+ struct page *page;
+ struct vvp_io *vio;
unsigned long ra_flags;
pgoff_t pg_offset;
int result;
goto out_err;
vio = vvp_env_io(env);
-
vio->u.fault.ft_vma = vma;
+ vio->u.fault.ft_vmpage = NULL;
vio->u.fault.nopage.ft_address = address;
vio->u.fault.nopage.ft_type = type;
result = cl_io_loop(env, io);
-out_err:
- if (result == 0) {
- LASSERT(io->u.ci_fault.ft_page != NULL);
- page = vio->u.fault.ft_vmpage;
- } else {
- if (result == -ENOMEM)
- page = NOPAGE_OOM;
+ page = vio->u.fault.ft_vmpage;
+ if (page != NULL) {
+ LASSERT(PageLocked(page));
+ unlock_page(page);
+
+ if (result != 0)
+ page_cache_release(page);
}
+ LASSERT(ergo(result == 0, io->u.ci_fault.ft_page != NULL));
+out_err:
+ if (result != 0)
+ page = result == -ENOMEM ? NOPAGE_OOM : NOPAGE_SIGBUS;
+
vma->vm_flags &= ~VM_RAND_READ;
vma->vm_flags |= ra_flags;
{
struct lu_env *env;
struct cl_io *io;
- struct vvp_io *vio = NULL;
+ struct vvp_io *vio;
unsigned long ra_flags;
struct cl_env_nest nest;
int result;
goto out_err;
vio = vvp_env_io(env);
-
vio->u.fault.ft_vma = vma;
+ vio->u.fault.ft_vmpage = NULL;
vio->u.fault.fault.ft_vmf = vmf;
result = cl_io_loop(env, io);
- fault_ret = vio->u.fault.fault.ft_flags;
+ if (unlikely(result != 0 && vio->u.fault.ft_vmpage != NULL)) {
+ struct page *vmpage = vio->u.fault.ft_vmpage;
+
+ LASSERT((vio->u.fault.fault.ft_flags & VM_FAULT_LOCKED) &&
+ PageLocked(vmpage));
+ unlock_page(vmpage);
+ page_cache_release(vmpage);
+ vmf->page = NULL;
+ }
+ fault_ret = vio->u.fault.fault.ft_flags;
out_err:
if (result != 0)
fault_ret |= VM_FAULT_ERROR;
LL_CDEBUG_PAGE(D_PAGE, vmpage, "got addr %lu type %lx\n",
cfio->nopage.ft_address, (long)cfio->nopage.ft_type);
+ lock_page(vmpage);
+ if (vmpage->mapping == NULL) {
+ CERROR("vmpage %lu@%p was truncated!\n", vmpage->index, vmpage);
+ unlock_page(vmpage);
+ page_cache_release(vmpage);
+ return -EFAULT;
+ }
+
cfio->ft_vmpage = vmpage;
return 0;
LL_CDEBUG_PAGE(D_PAGE, cfio->fault.ft_vmf->page,
"got addr %p type NOPAGE\n",
cfio->fault.ft_vmf->virtual_address);
- /*XXX workaround to bug in CLIO - he deadlocked with
- lock cancel if page locked */
- if (likely(cfio->fault.ft_flags & VM_FAULT_LOCKED)) {
- unlock_page(cfio->fault.ft_vmf->page);
- cfio->fault.ft_flags &= ~VM_FAULT_LOCKED;
- }
+ LASSERT(cfio->fault.ft_flags & VM_FAULT_LOCKED);
cfio->ft_vmpage = cfio->fault.ft_vmf->page;
return 0;
kernel_result = vvp_io_kernel_fault(cfio);
if (kernel_result != 0)
return kernel_result;
- /* Temporarily lock vmpage to keep cl_page_find() happy. */
- lock_page(cfio->ft_vmpage);
+
page = cl_page_find(env, obj, fio->ft_index, cfio->ft_vmpage,
CPT_CACHEABLE);
- unlock_page(cfio->ft_vmpage);
- if (IS_ERR(page)) {
- page_cache_release(cfio->ft_vmpage);
- cfio->ft_vmpage = NULL;
+ if (IS_ERR(page))
return PTR_ERR(page);
- }
size = i_size_read(inode);
last = cl_index(obj, size - 1);
hdr = cl_object_header(o);
cfs_atomic_inc(&site->cs_pages.cs_lookup);
- CDEBUG(D_PAGE, "%lu@"DFID" %p %lu %d\n",
+ CDEBUG(D_PAGE, "%lu@"DFID" %p %lx %d\n",
idx, PFID(&hdr->coh_lu.loh_fid), vmpage, vmpage->private, type);
/* fast path. */
if (type == CPT_CACHEABLE) {