* Author: Nikita Danilov <nikita.danilov@sun.com>
*/
-/** \addtogroup lov lov @{ */
-
#define DEBUG_SUBSYSTEM S_LOV
#include "lov_cl_internal.h"
+/** \addtogroup lov
+ * @{
+ */
+
/*****************************************************************************
*
* Layout operations.
if (r0->lo_sub != NULL) {
result = 0;
subconf->coc_inode = conf->coc_inode;
+ cfs_spin_lock_init(&r0->lo_sub_lock);
/*
* Create stripe cl_objects.
*/
cfs_waitlink_t *waiter;
r0 = &lov->u.raid0;
- sub = lovsub2cl(los);
LASSERT(r0->lo_sub[idx] == los);
+ sub = lovsub2cl(los);
+ site = sub->co_lu.lo_dev->ld_site;
+
cl_object_kill(env, sub);
/* release a reference to the sub-object and ... */
lu_object_ref_del(&sub->co_lu, "lov-parent", lov);
* ->lo_sub[] slot in lovsub_object_fini() */
if (r0->lo_sub[idx] == los) {
waiter = &lov_env_info(env)->lti_waiter;
- site = sub->co_lu.lo_dev->ld_site;
cfs_waitlink_init(waiter);
cfs_waitq_add(&site->ls_marche_funebre, waiter);
- set_current_state(CFS_TASK_UNINT);
-
- while (r0->lo_sub[idx] == los)
+ cfs_set_current_state(CFS_TASK_UNINT);
+ while (1) {
/* this wait-queue is signaled at the end of
* lu_object_free(). */
- cfs_waitq_wait(waiter, CFS_TASK_UNINT);
+ cfs_set_current_state(CFS_TASK_UNINT);
+ cfs_spin_lock(&r0->lo_sub_lock);
+ if (r0->lo_sub[idx] == los) {
+ cfs_spin_unlock(&r0->lo_sub_lock);
+ cfs_waitq_wait(waiter, CFS_TASK_UNINT);
+ } else {
+ cfs_spin_unlock(&r0->lo_sub_lock);
+ cfs_set_current_state(CFS_TASK_RUNNING);
+ break;
+ }
+ }
cfs_waitq_del(&site->ls_marche_funebre, waiter);
}
LASSERT(r0->lo_sub[idx] == NULL);
\
__lock &= __obj->lo_owner != cfs_current(); \
if (__lock) \
- down_read(&__obj->lo_type_guard); \
+ cfs_down_read(&__obj->lo_type_guard); \
__result = LOV_2DISPATCH_NOLOCK(obj, op, __VA_ARGS__); \
if (__lock) \
- up_read(&__obj->lo_type_guard); \
+ cfs_up_read(&__obj->lo_type_guard); \
__result; \
})
enum lov_layout_type __llt; \
\
if (__obj->lo_owner != cfs_current()) \
- down_read(&__obj->lo_type_guard); \
+ cfs_down_read(&__obj->lo_type_guard); \
__llt = __obj->lo_type; \
LASSERT(0 <= __llt && __llt < ARRAY_SIZE(lov_dispatch)); \
lov_dispatch[__llt].op(__VA_ARGS__); \
if (__obj->lo_owner != cfs_current()) \
- up_read(&__obj->lo_type_guard); \
+ cfs_up_read(&__obj->lo_type_guard); \
} while (0)
static int lov_layout_change(const struct lu_env *env,
cl_env_reexit(cookie);
old_ops->llo_fini(env, obj, &obj->u);
- LASSERT(list_empty(&hdr->coh_locks));
+ LASSERT(cfs_list_empty(&hdr->coh_locks));
LASSERT(hdr->coh_tree.rnode == NULL);
LASSERT(hdr->coh_pages == 0);
int result;
ENTRY;
- init_rwsem(&lov->lo_type_guard);
+ cfs_init_rwsem(&lov->lo_type_guard);
/* no locking is necessary, as object is being created */
lov->lo_type = cconf->u.coc_md->lsm != NULL ? LLT_RAID0 : LLT_EMPTY;
* Currently only LLT_EMPTY -> LLT_RAID0 transition is supported.
*/
LASSERT(lov->lo_owner != cfs_current());
- down_write(&lov->lo_type_guard);
+ cfs_down_write(&lov->lo_type_guard);
LASSERT(lov->lo_owner == NULL);
lov->lo_owner = cfs_current();
if (lov->lo_type == LLT_EMPTY && conf->u.coc_md->lsm != NULL)
else
result = -EOPNOTSUPP;
lov->lo_owner = NULL;
- up_write(&lov->lo_type_guard);
+ cfs_up_write(&lov->lo_type_guard);
RETURN(result);
}
};
struct lu_object *lov_object_alloc(const struct lu_env *env,
- const struct lu_object_header *_,
+ const struct lu_object_header *unused,
struct lu_device *dev)
{
struct lov_object *lov;