* GPL HEADER END
*/
/*
- * Copyright (c) 2012, 2014, Intel Corporation.
+ * Copyright (c) 2012, 2017, Intel Corporation.
* Use is subject to license terms.
*
* Author: Johann Lombardi <johann.lombardi@intel.com>
}
CDEBUG(D_QUOTA, "%s: global quota lock successfully acquired, glb "
- "fid:"DFID", glb ver:"LPU64", slv fid:"DFID", slv ver:"LPU64"\n",
+ "fid:"DFID", glb ver:%llu, slv fid:"DFID", slv ver:%llu\n",
qsd->qsd_svname, PFID(&req_qbody->qb_fid),
lvb->lvb_glb_ver, PFID(&rep_qbody->qb_slv_fid),
rep_qbody->qb_slv_ver);
rc = qsd_update_index(env, qqi, qid, global, 0, rec);
out:
+
+ if (global && qid->qid_uid == 0) {
+ struct lquota_glb_rec *glb_rec = (struct lquota_glb_rec *)rec;
+ qsd_update_default_quota(qqi, glb_rec->qbr_hardlimit,
+ glb_rec->qbr_softlimit,
+ glb_rec->qbr_time);
+ }
+
lqe_putref(lqe);
RETURN(rc);
}
/* let's do a 1MB bulk */
npages = min_t(unsigned int, OFD_MAX_BRW_SIZE, 1 << 20);
- npages /= PAGE_CACHE_SIZE;
+ npages /= PAGE_SIZE;
/* allocate pages for bulk index read */
OBD_ALLOC(pages, npages * sizeof(*pages));
ver = ii->ii_version;
pg_cnt = (ii->ii_count + (LU_PAGE_COUNT) - 1);
- pg_cnt >>= PAGE_CACHE_SHIFT - LU_PAGE_SHIFT;
+ pg_cnt >>= PAGE_SHIFT - LU_PAGE_SHIFT;
if (pg_cnt > npages) {
CERROR("%s: master returned more pages than expected, %u > %u"
if (rc == 0) {
rc = qsd_write_version(env, qqi, ver, global);
if (rc)
- CERROR("%s: write version "LPU64" to "DFID" failed. "
- "%d\n", qsd->qsd_svname, ver, PFID(fid), rc);
+ CERROR("%s: write version %llu to "DFID" failed : rc = %d\n",
+ qsd->qsd_svname, ver, PFID(fid), rc);
}
RETURN(rc);
struct qsd_qtype_info *qqi = (struct qsd_qtype_info *)args;
struct qsd_instance *qsd = qqi->qqi_qsd;
struct ptlrpc_thread *thread = &qqi->qqi_reint_thread;
- struct l_wait_info lwi = { 0 };
int rc;
ENTRY;
qti = qsd_info(env);
/* wait for the connection to master established */
- l_wait_event(thread->t_ctl_waitq,
- qsd_connected(qsd) || !thread_is_running(thread), &lwi);
+ wait_event_idle(thread->t_ctl_waitq,
+ qsd_connected(qsd) || !thread_is_running(thread));
/* Step 1: enqueue global index lock */
if (!thread_is_running(thread))
if (rc)
GOTO(out_env_init, rc);
- CDEBUG(D_QUOTA, "%s: glb_ver:"LPU64"/"LPU64",slv_ver:"LPU64"/"
- LPU64"\n", qsd->qsd_svname,
+ CDEBUG(D_QUOTA, "%s: glb_ver:%llu/%llu,slv_ver:%llu/"
+ "%llu\n", qsd->qsd_svname,
qti->qti_lvb.lvb_glb_ver, qqi->qqi_glb_ver,
qti->qti_slv_ver, qqi->qqi_slv_ver);
}
}
/* wait for the qsd instance started (target recovery done) */
- l_wait_event(thread->t_ctl_waitq,
- qsd_started(qsd) || !thread_is_running(thread), &lwi);
+ wait_event_idle(thread->t_ctl_waitq,
+ qsd_started(qsd) || !thread_is_running(thread));
if (!thread_is_running(thread))
GOTO(out_lock, rc = 0);
void qsd_stop_reint_thread(struct qsd_qtype_info *qqi)
{
struct ptlrpc_thread *thread = &qqi->qqi_reint_thread;
- struct l_wait_info lwi = { 0 };
if (!thread_is_stopped(thread)) {
thread_set_flags(thread, SVC_STOPPING);
wake_up(&thread->t_ctl_waitq);
- l_wait_event(thread->t_ctl_waitq,
- thread_is_stopped(thread), &lwi);
+ wait_event_idle(thread->t_ctl_waitq,
+ thread_is_stopped(thread));
}
}
{
struct ptlrpc_thread *thread = &qqi->qqi_reint_thread;
struct qsd_instance *qsd = qqi->qqi_qsd;
- struct l_wait_info lwi = { 0 };
struct task_struct *task;
int rc;
char *name;
ENTRY;
+ /* do not try to start a new thread as this can lead to a deadlock */
+ if (current->flags & (PF_MEMALLOC | PF_KSWAPD))
+ RETURN(0);
+
+ if (qsd->qsd_dev->dd_rdonly)
+ RETURN(0);
+
/* don't bother to do reintegration when quota isn't enabled */
if (!qsd_type_enabled(qsd, qqi->qqi_qtype))
RETURN(0);
- if (qsd->qsd_acct_failed)
+ if (qqi->qqi_acct_failed)
/* no space accounting support, can't enable enforcement */
RETURN(0);
RETURN(rc);
}
- l_wait_event(thread->t_ctl_waitq,
- thread_is_running(thread) || thread_is_stopped(thread),
- &lwi);
+ wait_event_idle(thread->t_ctl_waitq,
+ thread_is_running(thread) || thread_is_stopped(thread));
RETURN(0);
}