void init_waitqueue_entry_current(wait_queue_t *link)
{
- cfs_task_t * task = cfs_current();
+ struct task_struct * task = current;
PTASK_SLOT slot = NULL;
if (!task) {
link->event = &(slot->Event);
link->hits = &(slot->hits);
- cfs_atomic_inc(&slot->count);
+ atomic_inc(&slot->count);
CFS_INIT_LIST_HEAD(&(link->waitq[0].link));
CFS_INIT_LIST_HEAD(&(link->waitq[1].link));
void cfs_waitlink_fini(wait_queue_t *link)
{
- cfs_task_t * task = cfs_current();
+ struct task_struct * task = current;
PTASK_SLOT slot = NULL;
if (!task) {
cfs_assert(link->waitq[0].waitq == NULL);
cfs_assert(link->waitq[1].waitq == NULL);
- cfs_atomic_dec(&slot->count);
+ atomic_dec(&slot->count);
}
LASSERT( result == FALSE || result == TRUE );
if (result) {
- cfs_atomic_inc(waitl->hits);
+ atomic_inc(waitl->hits);
}
if ((waitl->flags & CFS_WAITQ_EXCLUSIVE) && --nr == 0)
LASSERT(link != NULL);
LASSERT(link->magic == CFS_WAITLINK_MAGIC);
- if (cfs_atomic_read(link->hits) > 0) {
- cfs_atomic_dec(link->hits);
- LASSERT((__u32)cfs_atomic_read(link->hits) < (__u32)0xFFFFFF00);
+ if (atomic_read(link->hits) > 0) {
+ atomic_dec(link->hits);
+ LASSERT((__u32)atomic_read(link->hits) < (__u32)0xFFFFFF00);
} else {
cfs_wait_event_internal(link->event, 0);
}
int64_t timeout)
{
- if (cfs_atomic_read(link->hits) > 0) {
- cfs_atomic_dec(link->hits);
- LASSERT((__u32)cfs_atomic_read(link->hits) < (__u32)0xFFFFFF00);
+ if (atomic_read(link->hits) > 0) {
+ atomic_dec(link->hits);
+ LASSERT((__u32)atomic_read(link->hits) < (__u32)0xFFFFFF00);
return (int64_t)TRUE;
}