* GPL HEADER END
*/
/*
- * Copyright (c) 2011, 2012, Intel, Inc.
+ * Copyright (c) 2012, 2014, Intel Corporation.
* Use is subject to license terms.
*
* Author: Johann Lombardi <johann.lombardi@intel.com>
* Author: Niu Yawei <yawei.niu@intel.com>
*/
-#ifndef EXPORT_SYMTAB
-# define EXPORT_SYMTAB
-#endif
-
#define DEBUG_SUBSYSTEM S_LQUOTA
+#include <linux/kthread.h>
#include "qsd_internal.h"
/*
struct quota_body *req_qbody,
struct quota_body *rep_qbody,
struct lustre_handle *lockh,
- union ldlm_wire_lvb *lvb,
+ struct lquota_lvb *lvb,
void *arg, int rc)
{
struct qsd_instance *qsd = qqi->qqi_qsd;
CDEBUG(D_QUOTA, "%s: global quota lock successfully acquired, glb "
"fid:"DFID", glb ver:"LPU64", slv fid:"DFID", slv ver:"LPU64"\n",
qsd->qsd_svname, PFID(&req_qbody->qb_fid),
- lvb->l_lquota.lvb_glb_ver, PFID(&rep_qbody->qb_slv_fid),
+ lvb->lvb_glb_ver, PFID(&rep_qbody->qb_slv_fid),
rep_qbody->qb_slv_ver);
*slv_ver = rep_qbody->qb_slv_ver;
static int qsd_reint_entries(const struct lu_env *env,
struct qsd_qtype_info *qqi,
struct idx_info *ii, bool global,
- cfs_page_t **pages,
+ struct page **pages,
unsigned int npages, bool need_swab)
{
struct qsd_thread_info *qti = qsd_info(env);
size = ii->ii_recsize + ii->ii_keysize;
for (i = 0; i < npages; i++) {
- union lu_page *lip = cfs_kmap(pages[i]);
+ union lu_page *lip = kmap(pages[i]);
for (j = 0; j < LU_PAGE_COUNT; j++) {
if (need_swab)
lip++;
}
out:
- cfs_kunmap(pages[i]);
+ kunmap(pages[i]);
if (rc)
break;
}
struct qsd_instance *qsd = qqi->qqi_qsd;
struct idx_info *ii = &qti->qti_ii;
struct lu_fid *fid;
- cfs_page_t **pages = NULL;
+ struct page **pages = NULL;
unsigned int npages, pg_cnt;
__u64 start_hash = 0, ver = 0;
bool need_swab = false;
fid = global ? &qqi->qqi_fid : &qqi->qqi_slv_fid;
/* let's do a 1MB bulk */
- npages = min_t(unsigned int, PTLRPC_MAX_BRW_SIZE, 1 << 20);
- npages /= CFS_PAGE_SIZE;
+ npages = min_t(unsigned int, OFD_MAX_BRW_SIZE, 1 << 20);
+ npages /= PAGE_CACHE_SIZE;
/* allocate pages for bulk index read */
OBD_ALLOC(pages, npages * sizeof(*pages));
if (pages == NULL)
GOTO(out, rc = -ENOMEM);
for (i = 0; i < npages; i++) {
- pages[i] = cfs_alloc_page(CFS_ALLOC_STD);
+ pages[i] = alloc_page(GFP_IOFS);
if (pages[i] == NULL)
GOTO(out, rc = -ENOMEM);
}
ver = ii->ii_version;
pg_cnt = (ii->ii_count + (LU_PAGE_COUNT) - 1);
- pg_cnt >>= CFS_PAGE_SHIFT - LU_PAGE_SHIFT;
+ pg_cnt >>= PAGE_CACHE_SHIFT - LU_PAGE_SHIFT;
if (pg_cnt > npages) {
CERROR("%s: master returned more pages than expected, %u > %u"
if (pages != NULL) {
for (i = 0; i < npages; i++)
if (pages[i] != NULL)
- cfs_free_page(pages[i]);
+ __free_page(pages[i]);
OBD_FREE(pages, npages * sizeof(*pages));
}
LASSERT(qqi->qqi_glb_obj != NULL);
iops = &qqi->qqi_glb_obj->do_index_ops->dio_it;
- it = iops->init(env, qqi->qqi_glb_obj, 0, BYPASS_CAPA);
+ it = iops->init(env, qqi->qqi_glb_obj, 0);
if (IS_ERR(it)) {
CWARN("%s: Initialize it for "DFID" failed. %ld\n",
qsd->qsd_svname, PFID(&qqi->qqi_fid), PTR_ERR(it));
int rc;
ENTRY;
- cfs_daemonize("qsd_reint");
-
CDEBUG(D_QUOTA, "%s: Starting reintegration thread for "DFID"\n",
qsd->qsd_svname, PFID(&qqi->qqi_fid));
lu_ref_add(&qqi->qqi_reference, "reint_thread", thread);
thread_set_flags(thread, SVC_RUNNING);
- cfs_waitq_signal(&thread->t_ctl_waitq);
+ wake_up(&thread->t_ctl_waitq);
OBD_ALLOC_PTR(env);
if (env == NULL)
ldlm_lock_addref_try(&qqi->qqi_lockh, qsd_glb_einfo.ei_mode) == 0) {
read_unlock(&qsd->qsd_lock);
/* force refresh of global & slave index copy */
- qti->qti_lvb.l_lquota.lvb_glb_ver = ~0ULL;
+ qti->qti_lvb.lvb_glb_ver = ~0ULL;
qti->qti_slv_ver = ~0ULL;
} else {
/* no valid lock found, let's enqueue a new one */
CDEBUG(D_QUOTA, "%s: glb_ver:"LPU64"/"LPU64",slv_ver:"LPU64"/"
LPU64"\n", qsd->qsd_svname,
- qti->qti_lvb.l_lquota.lvb_glb_ver, qqi->qqi_glb_ver,
+ qti->qti_lvb.lvb_glb_ver, qqi->qqi_glb_ver,
qti->qti_slv_ver, qqi->qqi_slv_ver);
}
OBD_FAIL_TIMEOUT(OBD_FAIL_QUOTA_DELAY_REINT, 10);
- if (qqi->qqi_glb_ver != qti->qti_lvb.l_lquota.lvb_glb_ver) {
+ if (qqi->qqi_glb_ver != qti->qti_lvb.lvb_glb_ver) {
rc = qsd_reint_index(env, qqi, true);
if (rc) {
CWARN("%s: reint global for "DFID" failed. %d\n",
if (qqi->qqi_slv_ver != qti->qti_slv_ver) {
rc = qsd_reint_index(env, qqi, false);
if (rc) {
- CWARN("%s: Reint slave for "DFID" failed. %d\n",
+ CWARN("%s: reintegration for "DFID" failed with %d\n",
qsd->qsd_svname, PFID(&qqi->qqi_slv_fid), rc);
GOTO(out_lock, rc);
}
qsd_bump_version(qqi, qqi->qqi_slv_ver, false);
}
- /* wait for the connection to master established */
+ /* wait for the qsd instance started (target recovery done) */
l_wait_event(thread->t_ctl_waitq,
qsd_started(qsd) || !thread_is_running(thread), &lwi);
/* Step 4: start reconciliation for each enforced ID */
rc = qsd_reconciliation(env, qqi);
if (rc)
- CWARN("%s: reconciliation failed. "DFID", %d\n",
- qsd->qsd_svname, PFID(&qti->qti_fid), rc);
+ CWARN("%s: reconciliation for "DFID" failed with %d\n",
+ qsd->qsd_svname, PFID(&qqi->qqi_slv_fid), rc);
EXIT;
out_lock:
lu_ref_del(&qqi->qqi_reference, "reint_thread", thread);
thread_set_flags(thread, SVC_STOPPED);
- cfs_waitq_signal(&thread->t_ctl_waitq);
+ wake_up(&thread->t_ctl_waitq);
return rc;
}
if (!thread_is_stopped(thread)) {
thread_set_flags(thread, SVC_STOPPING);
- cfs_waitq_signal(&thread->t_ctl_waitq);
+ wake_up(&thread->t_ctl_waitq);
l_wait_event(thread->t_ctl_waitq,
thread_is_stopped(thread), &lwi);
}
}
-static int qsd_entry_iter_cb(cfs_hash_t *hs, cfs_hash_bd_t *bd,
- cfs_hlist_node_t *hnode, void *data)
+static int qsd_entry_iter_cb(struct cfs_hash *hs, struct cfs_hash_bd *bd,
+ struct hlist_node *hnode, void *data)
{
struct lquota_entry *lqe;
int *pending = (int *)data;
- lqe = cfs_hlist_entry(hnode, struct lquota_entry, lqe_hash);
+ lqe = hlist_entry(hnode, struct lquota_entry, lqe_hash);
LASSERT(atomic_read(&lqe->lqe_ref) > 0);
lqe_read_lock(lqe);
/* any pending quota adjust? */
spin_lock(&qsd->qsd_adjust_lock);
- cfs_list_for_each_entry_safe(lqe, n, &qsd->qsd_adjust_list, lqe_link) {
+ list_for_each_entry_safe(lqe, n, &qsd->qsd_adjust_list, lqe_link) {
if (lqe2qqi(lqe) == qqi) {
- cfs_list_del_init(&lqe->lqe_link);
+ list_del_init(&lqe->lqe_link);
lqe_putref(lqe);
}
}
/* any pending updates? */
read_lock(&qsd->qsd_lock);
- cfs_list_for_each_entry(upd, &qsd->qsd_upd_list, qur_link) {
+ list_for_each_entry(upd, &qsd->qsd_upd_list, qur_link) {
if (upd->qur_qqi == qqi) {
read_unlock(&qsd->qsd_lock);
CDEBUG(D_QUOTA, "%s: pending %s updates for type:%d.\n",
{
struct ptlrpc_thread *thread = &qqi->qqi_reint_thread;
struct qsd_instance *qsd = qqi->qqi_qsd;
- struct l_wait_info lwi = { 0 };
- int rc;
+ struct l_wait_info lwi = { 0 };
+ struct task_struct *task;
+ int rc;
+ char *name;
ENTRY;
/* don't bother to do reintegration when quota isn't enabled */
- if (!qsd_type_enabled(qqi->qqi_qsd, qqi->qqi_qtype))
+ if (!qsd_type_enabled(qsd, qqi->qqi_qtype))
+ RETURN(0);
+
+ if (qsd->qsd_acct_failed)
+ /* no space accounting support, can't enable enforcement */
RETURN(0);
/* check if the reintegration has already started or finished */
RETURN(0);
}
- rc = cfs_create_thread(qsd_reint_main, (void *)qqi, 0);
- if (rc < 0) {
+ OBD_ALLOC(name, MTI_NAME_MAXLEN);
+ if (name == NULL)
+ RETURN(-ENOMEM);
+
+ snprintf(name, MTI_NAME_MAXLEN, "qsd_reint_%d.%s",
+ qqi->qqi_qtype, qsd->qsd_svname);
+
+ task = kthread_run(qsd_reint_main, qqi, name);
+ OBD_FREE(name, MTI_NAME_MAXLEN);
+
+ if (IS_ERR(task)) {
+ rc = PTR_ERR(task);
thread_set_flags(thread, SVC_STOPPED);
write_lock(&qsd->qsd_lock);
qqi->qqi_reint = 0;