#define RQ_LATER 0x4
#define RQ_STOP 0x8
static int rq_state = 0;
-static cfs_waitq_t rq_waitq;
+static wait_queue_head_t rq_waitq;
static DECLARE_COMPLETION(rq_exit);
static void do_requeue(struct config_llog_data *cld)
} else {
rq_state |= RQ_NOW;
spin_unlock(&config_list_lock);
- cfs_waitq_signal(&rq_waitq);
+ wake_up(&rq_waitq);
}
EXIT;
}
rq_state |= RQ_STOP;
spin_unlock(&config_list_lock);
if (running) {
- cfs_waitq_signal(&rq_waitq);
+ wake_up(&rq_waitq);
wait_for_completion(&rq_exit);
}
}
if (cfs_atomic_inc_return(&mgc_count) == 1) {
rq_state = 0;
- cfs_waitq_init(&rq_waitq);
+ init_waitqueue_head(&rq_waitq);
/* start requeue thread */
rc = PTR_ERR(kthread_run(mgc_requeue_thread, NULL,
spin_lock(&config_list_lock);
rq_state |= RQ_NOW;
spin_unlock(&config_list_lock);
- cfs_waitq_signal(&rq_waitq);
+ wake_up(&rq_waitq);
/* TODO: Help the MGS rebuild nidtbl. -jay */
}