env = cl_env_get(&refcheck);
LASSERT(!IS_ERR(env));
- io = vvp_env_io(env)->cui_cl.cis_io;
+ io = vvp_env_io(env)->vui_cl.cis_io;
LASSERT(io != NULL);
for (seg = 0; seg < nr_segs; seg++) {
}
out:
if (tot_bytes > 0) {
- struct vvp_io *cio = vvp_env_io(env);
+ struct vvp_io *vio = vvp_env_io(env);
/* no commit async for direct IO */
- cio->u.write.cui_written += tot_bytes;
+ vio->u.write.vui_written += tot_bytes;
}
cl_env_put(env, &refcheck);
if (unlikely(vmpage == NULL ||
PageDirty(vmpage) || PageWriteback(vmpage))) {
- struct vvp_io *cio = vvp_env_io(env);
- struct cl_page_list *plist = &cio->u.write.cui_queue;
+ struct vvp_io *vio = vvp_env_io(env);
+ struct cl_page_list *plist = &vio->u.write.vui_queue;
/* if the page is already in dirty cache, we have to commit
* the pages right now; otherwise, it may cause deadlock
struct ll_cl_context *lcc = fsdata;
const struct lu_env *env;
struct cl_io *io;
- struct vvp_io *cio;
+ struct vvp_io *vio;
struct cl_page *page;
unsigned from = pos & (PAGE_CACHE_SIZE - 1);
bool unplug = false;
env = lcc->lcc_env;
page = lcc->lcc_page;
io = lcc->lcc_io;
- cio = vvp_env_io(env);
+ vio = vvp_env_io(env);
LASSERT(cl_page_is_owned(page, io));
if (copied > 0) {
- struct cl_page_list *plist = &cio->u.write.cui_queue;
+ struct cl_page_list *plist = &vio->u.write.vui_queue;
lcc->lcc_page = NULL; /* page will be queued */
/* Add it into write queue */
cl_page_list_add(plist, page);
if (plist->pl_nr == 1) /* first page */
- cio->u.write.cui_from = from;
+ vio->u.write.vui_from = from;
else
LASSERT(from == 0);
- cio->u.write.cui_to = from + copied;
+ vio->u.write.vui_to = from + copied;
/* To address the deadlock in balance_dirty_pages() where
* this dirty page may be written back in the same thread. */