- cfs_list_for_each_entry_safe_reverse(tmp, n, list, qur_link) {
+ list_for_each_entry_safe_reverse(tmp, n, list, qur_link) {
/* There could be some legacy records which have duplicated
* version. Imagine following scenario: slave received global
/* There could be some legacy records which have duplicated
* version. Imagine following scenario: slave received global
LASSERT(tmp->qur_lqe);
LQUOTA_ERROR(tmp->qur_lqe, "Found a conflict record "
"with ver:"LPU64"", tmp->qur_ver);
LASSERT(tmp->qur_lqe);
LQUOTA_ERROR(tmp->qur_lqe, "Found a conflict record "
"with ver:"LPU64"", tmp->qur_ver);
-static void qsd_kickoff_deferred(struct qsd_qtype_info *qqi, cfs_list_t *list,
- __u64 ver)
+static void qsd_kickoff_deferred(struct qsd_qtype_info *qqi,
+ struct list_head *list, __u64 ver)
/* Get the first update record in the list, which has the smallest
* version, discard all records with versions smaller than the current
* one */
/* Get the first update record in the list, which has the smallest
* version, discard all records with versions smaller than the current
* one */
- cfs_list_for_each_entry_safe(upd, tmp, list, qur_link) {
+ list_for_each_entry_safe(upd, tmp, list, qur_link) {
CDEBUG(D_QUOTA, "%s: skipping deferred update ver:"
LPU64"/"LPU64", global:%d, qid:"LPU64"\n",
qqi->qqi_qsd->qsd_svname, upd->qur_ver, ver,
CDEBUG(D_QUOTA, "%s: skipping deferred update ver:"
LPU64"/"LPU64", global:%d, qid:"LPU64"\n",
qqi->qqi_qsd->qsd_svname, upd->qur_ver, ver,
idx_ver = global ? &qqi->qqi_glb_ver : &qqi->qqi_slv_ver;
list = global ? &qqi->qqi_deferred_glb : &qqi->qqi_deferred_slv;
idx_ver = global ? &qqi->qqi_glb_ver : &qqi->qqi_slv_ver;
list = global ? &qqi->qqi_deferred_glb : &qqi->qqi_deferred_slv;
/* Out of order update (the one with smaller version hasn't
* reached slave or hasn't been flushed to disk yet), or
* the reintegration is in progress. Defer the update. */
/* Out of order update (the one with smaller version hasn't
* reached slave or hasn't been flushed to disk yet), or
* the reintegration is in progress. Defer the update. */
/* the lqe is being queued for the per-ID lock cancel, we should
* cancel the lock cancel and re-add it for quota adjust */
/* the lqe is being queued for the per-ID lock cancel, we should
* cancel the lock cancel and re-add it for quota adjust */
struct lquota_entry, lqe_link);
if (cfs_time_beforeq_64(lqe->lqe_adjust_time,
cfs_time_current_64()))
struct lquota_entry, lqe_link);
if (cfs_time_beforeq_64(lqe->lqe_adjust_time,
cfs_time_current_64()))
- if (!cfs_list_empty(&qsd->qsd_upd_list)) {
- cfs_list_splice_init(&qsd->qsd_upd_list, upd);
+ if (!list_empty(&qsd->qsd_upd_list)) {
+ list_splice_init(&qsd->qsd_upd_list, upd);
struct qsd_instance *qsd = (struct qsd_instance *)arg;
struct ptlrpc_thread *thread = &qsd->qsd_upd_thread;
struct l_wait_info lwi;
struct qsd_instance *qsd = (struct qsd_instance *)arg;
struct ptlrpc_thread *thread = &qsd->qsd_upd_thread;
struct l_wait_info lwi;
thread_set_flags(thread, SVC_RUNNING);
wake_up(&thread->t_ctl_waitq);
thread_set_flags(thread, SVC_RUNNING);
wake_up(&thread->t_ctl_waitq);
lwi = LWI_TIMEOUT(cfs_time_seconds(QSD_WB_INTERVAL), NULL, NULL);
while (1) {
l_wait_event(thread->t_ctl_waitq,
qsd_job_pending(qsd, &queue, &uptodate) ||
!thread_is_running(thread), &lwi);
lwi = LWI_TIMEOUT(cfs_time_seconds(QSD_WB_INTERVAL), NULL, NULL);
while (1) {
l_wait_event(thread->t_ctl_waitq,
qsd_job_pending(qsd, &queue, &uptodate) ||
!thread_is_running(thread), &lwi);
- cfs_list_for_each_entry_safe(upd, n, &queue, qur_link) {
- cfs_list_del_init(&upd->qur_link);
+ list_for_each_entry_safe(upd, n, &queue, qur_link) {
+ list_del_init(&upd->qur_link);
qsd_process_upd(env, upd);
qsd_upd_free(upd);
}
spin_lock(&qsd->qsd_adjust_lock);
cur_time = cfs_time_current_64();
qsd_process_upd(env, upd);
qsd_upd_free(upd);
}
spin_lock(&qsd->qsd_adjust_lock);
cur_time = cfs_time_current_64();
- cfs_list_for_each_entry_safe(lqe, tmp, &qsd->qsd_adjust_list,
- lqe_link) {
+ while (!list_empty(&qsd->qsd_adjust_list)) {
+ lqe = list_entry(qsd->qsd_adjust_list.next,
+ struct lquota_entry, lqe_link);
/* deferred items are sorted by time */
if (!cfs_time_beforeq_64(lqe->lqe_adjust_time,
cur_time))
break;
/* deferred items are sorted by time */
if (!cfs_time_beforeq_64(lqe->lqe_adjust_time,
cur_time))
break;
ENTRY;
task = kthread_run(qsd_upd_thread, (void *)qsd,
"lquota_wb_%s", qsd->qsd_svname);
if (IS_ERR(task)) {
ENTRY;
task = kthread_run(qsd_upd_thread, (void *)qsd,
"lquota_wb_%s", qsd->qsd_svname);
if (IS_ERR(task)) {
- cfs_list_for_each_entry_safe(upd, tmp, &qqi->qqi_deferred_glb,
- qur_link) {
+ list_for_each_entry_safe(upd, tmp, &qqi->qqi_deferred_glb,
+ qur_link) {
CWARN("%s: Free global deferred upd: ID:"LPU64", "
"ver:"LPU64"/"LPU64"\n", qsd->qsd_svname,
upd->qur_qid.qid_uid, upd->qur_ver,
CWARN("%s: Free global deferred upd: ID:"LPU64", "
"ver:"LPU64"/"LPU64"\n", qsd->qsd_svname,
upd->qur_qid.qid_uid, upd->qur_ver,
- cfs_list_for_each_entry_safe(upd, tmp, &qqi->qqi_deferred_slv,
- qur_link) {
+ list_for_each_entry_safe(upd, tmp, &qqi->qqi_deferred_slv,
+ qur_link) {
CWARN("%s: Free slave deferred upd: ID:"LPU64", "
"ver:"LPU64"/"LPU64"\n", qsd->qsd_svname,
upd->qur_qid.qid_uid, upd->qur_ver,
CWARN("%s: Free slave deferred upd: ID:"LPU64", "
"ver:"LPU64"/"LPU64"\n", qsd->qsd_svname,
upd->qur_qid.qid_uid, upd->qur_ver,
- while (!cfs_list_empty(&qsd->qsd_adjust_list)) {
- lqe = cfs_list_entry(qsd->qsd_adjust_list.next,
- struct lquota_entry, lqe_link);
- cfs_list_del_init(&lqe->lqe_link);
+ while (!list_empty(&qsd->qsd_adjust_list)) {
+ lqe = list_entry(qsd->qsd_adjust_list.next,
+ struct lquota_entry, lqe_link);
+ list_del_init(&lqe->lqe_link);