struct cl_2queue *queue = NULL;
struct cl_sync_io *anchor = NULL;
bool holdinglock = false;
- bool lockedbymyself = true;
int rc;
ENTRY;
if (!PageUptodate(vmpage) && !PageDirty(vmpage) &&
!PageWriteback(vmpage)) {
/* read page */
- /* set PagePrivate2 to detect special case of empty page
- * in osc_brw_fini_request()
+ /* Set PagePrivate2 to detect special case of empty page
+ * in osc_brw_fini_request().
+ * It is also used to tell ll_io_read_page() that we do not
+ * want the vmpage to be unlocked.
*/
SetPagePrivate2(vmpage);
rc = ll_io_read_page(env, io, clpage, NULL);
- if (!PagePrivate2(vmpage))
+ if (!PagePrivate2(vmpage)) {
/* PagePrivate2 was cleared in osc_brw_fini_request()
* meaning we read an empty page. In this case, in order
* to avoid allocating unnecessary block in truncated
* file, we must not zero and write as below. Subsequent
* server-side truncate will handle things correctly.
*/
+ cl_page_unassume(env, io, clpage);
GOTO(clpfini, rc = 0);
+ }
ClearPagePrivate2(vmpage);
if (rc)
GOTO(clpfini, rc);
- lockedbymyself = trylock_page(vmpage);
- cl_page_assume(env, io, clpage);
}
- /* zero range in page */
+ /* Thanks to PagePrivate2 flag, ll_io_read_page() did not unlock
+ * the vmpage, so we are good to proceed and zero range in page.
+ */
zero_user(vmpage, offset, len);
if (holdinglock && clpage) {
if (clpage)
cl_page_put(env, clpage);
pagefini:
- if (lockedbymyself) {
- unlock_page(vmpage);
- put_page(vmpage);
- }
+ unlock_page(vmpage);
+ put_page(vmpage);
rellock:
if (holdinglock)
cl_lock_release(env, lock);
pgoff_t ra_start_index = 0;
pgoff_t io_start_index;
pgoff_t io_end_index;
+ bool unlockpage = true;
ENTRY;
if (file) {
ras = &fd->fd_ras;
}
+ /* PagePrivate2 is set in ll_io_zero_page() to tell us the vmpage
+ * must not be unlocked after processing.
+ */
+ if (page->cp_vmpage && PagePrivate2(page->cp_vmpage))
+ unlockpage = false;
+
vpg = cl2vvp_page(cl_object_page_slice(page->cp_obj, page));
uptodate = vpg->vpg_defer_uptodate;
* route is implemented */
cl_page_discard(env, io, page);
}
- cl_page_disown(env, io, page);
+ if (unlockpage)
+ cl_page_disown(env, io, page);
}
/* TODO: discard all pages until page reinit route is implemented */