* GPL HEADER END
*/
/*
- * Copyright 2008 Sun Microsystems, Inc. All rights reserved
+ * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved.
* Use is subject to license terms.
*/
/*
#include <cl_object.h>
#include "lov_internal.h"
-/** \addtogroup lov lov @{ */
-
/** \defgroup lov lov
* Logical object volume layer. This layer implements data striping (raid0).
*
* cl_lock::cll_guard, and will be automatically cleared by the sub-lock
* when the latter is destroyed. When a sub-lock is canceled, a
* reference to it is removed from the top-lock array, and top-lock is
- * moved into CLS_NEW state. It is guaranteed that all sub-locks exits
+ * moved into CLS_NEW state. It is guaranteed that all sub-locks exist
* while their top-lock is in CLS_HELD or CLS_CACHED states.
*
* - IO's are not reference counted.
*
* To implement a connection between top and sub entities, lov layer is split
* into two pieces: lov ("upper half"), and lovsub ("bottom half"), both
- * implementing full set of cl-interfaces. For example, top-object has clu and
+ * implementing full set of cl-interfaces. For example, top-object has vvp and
* lov layers, and it's sub-object has lovsub and osc layers. lovsub layer is
* used to track child-parent relationship.
*
* Serializes access to lov_device::ld_emrg in low-memory
* conditions.
*/
- struct mutex ld_mutex;
+ cfs_mutex_t ld_mutex;
};
/**
LLT_EMPTY,
/** striped file */
LLT_RAID0,
- /** join file */
- LLT_JOIN,
LLT_NR
};
*
* \see lov_object::lo_type
*/
- struct rw_semaphore lo_type_guard;
+ cfs_rw_semaphore_t lo_type_guard;
/**
* Type of an object. Protected by lov_object::lo_type_guard.
*/
*/
struct lovsub_object **lo_sub;
/**
+ * protect lo_sub
+ */
+ cfs_spinlock_t lo_sub_lock;
+ /**
* When this is true, lov_object::lo_attr contains
* valid up to date attributes for a top-level
* object. This field is reset to 0 when attributes of
} raid0;
struct lov_layout_state_empty {
} empty;
- struct lov_layout_state_join {
- } join;
} u;
/**
* Thread that acquired lov_object::lo_type_guard in an exclusive
unsigned lls_nr_filled;
/**
* Set when sub-lock was canceled, while top-lock was being
- * unlocked.
+ * used, or unused.
*/
- int lls_unuse_race;
+ int lls_cancel_race:1;
/**
* An array of sub-locks
*
* A linkage into per sub-lock list of all corresponding top-locks,
* hanging off lovsub_lock::lss_parents.
*/
- struct list_head lll_list;
+ cfs_list_t lll_list;
};
/**
* List of top-locks that have given sub-lock as their part. Protected
* by cl_lock::cll_guard mutex.
*/
- struct list_head lss_parents;
+ cfs_list_t lss_parents;
/**
* Top-lock that initiated current operation on this sub-lock. This is
* only set during top-to-bottom lock operations like enqueue, and is
* Linkage into a list (hanging off lov_io::lis_active) of all
* sub-io's active for the current IO iteration.
*/
- struct list_head sub_linkage;
+ cfs_list_t sub_linkage;
/**
* true, iff cl_io_init() was successfully executed against
* lov_io_sub::sub_io.
/**
* List of active sub-io's.
*/
- struct list_head lis_active;
+ cfs_list_t lis_active;
};
struct lov_session {
LASSERT(lov->lo_type == LLT_RAID0);
raid0 = &lov->u.raid0;
- LASSERT(raid0->lo_lsm->lsm_wire.lw_magic == LOV_MAGIC);
+ LASSERT(raid0->lo_lsm->lsm_wire.lw_magic == LOV_MAGIC ||
+ raid0->lo_lsm->lsm_wire.lw_magic == LOV_MAGIC_V3);
return raid0;
}