ltd->ltd_refcount--;
if (ltd->ltd_refcount == 0 && ltd->ltd_death_row) {
struct lod_tgt_desc *tgt_desc, *tmp;
- int idx;
- CFS_LIST_HEAD(kill);
+ struct list_head kill;
+ unsigned int idx;
CDEBUG(D_CONFIG, "destroying %d ltd desc\n",
ltd->ltd_death_row);
+ INIT_LIST_HEAD(&kill);
+
cfs_foreach_bit(ltd->ltd_tgt_bitmap, idx) {
tgt_desc = LTD_TGT(ltd, idx);
LASSERT(tgt_desc);
if (!tgt_desc->ltd_reap)
continue;
- cfs_list_add(&tgt_desc->ltd_kill, &kill);
+ list_add(&tgt_desc->ltd_kill, &kill);
LTD_TGT(ltd, idx) = NULL;
/*FIXME: only support ost pool for now */
if (ltd == &lod->lod_ost_descs) {
mutex_unlock(<d->ltd_mutex);
up_read(<d->ltd_rw_sem);
- cfs_list_for_each_entry_safe(tgt_desc, tmp, &kill, ltd_kill) {
+ list_for_each_entry_safe(tgt_desc, tmp, &kill, ltd_kill) {
int rc;
- cfs_list_del(&tgt_desc->ltd_kill);
+ list_del(&tgt_desc->ltd_kill);
if (ltd == &lod->lod_ost_descs) {
/* remove from QoS structures */
rc = qos_del_tgt(lod, tgt_desc);
int lod_fini_tgt(const struct lu_env *env, struct lod_device *lod,
struct lod_tgt_descs *ltd, bool for_ost)
{
- int idx;
+ unsigned int idx;
if (ltd->ltd_tgts_size <= 0)
return 0;
return(rc);
}
-int lod_ea_store_resize(struct lod_thread_info *info, int size)
+int lod_ea_store_resize(struct lod_thread_info *info, size_t size)
{
- int round = size_roundup_power2(size);
+ __u32 round = size_roundup_power2(size);
LASSERT(round <=
lov_mds_md_size(LOV_MAX_STRIPE_COUNT, LOV_MAGIC_V3));
struct lov_mds_md_v1 *lmm;
struct lov_ost_data_v1 *objs;
__u32 magic;
- int i, rc, lmm_size;
- int cplen = 0;
+ int i, rc;
+ size_t lmm_size;
ENTRY;
LASSERT(lo);
objs = &lmm->lmm_objects[0];
} else {
struct lov_mds_md_v3 *v3 = (struct lov_mds_md_v3 *) lmm;
- cplen = strlcpy(v3->lmm_pool_name, lo->ldo_pool,
+ size_t cplen = strlcpy(v3->lmm_pool_name, lo->ldo_pool,
sizeof(v3->lmm_pool_name));
if (cplen >= sizeof(v3->lmm_pool_name))
RETURN(-E2BIG);
RETURN(rc);
}
-static int validate_lod_and_idx(struct lod_device *md, int idx)
+static int validate_lod_and_idx(struct lod_device *md, __u32 idx)
{
if (unlikely(idx >= md->lod_ost_descs.ltd_tgts_size ||
!cfs_bitmap_check(md->lod_ost_bitmap, idx))) {
struct lu_device *nd;
struct dt_object **stripe;
int stripe_len;
- int i, idx, rc = 0;
+ int i, rc = 0;
+ __u32 idx;
ENTRY;
LASSERT(lo != NULL);
rc = lod_parse_striping(env, lo, buf);
} else if (S_ISDIR(lu_object_attr(lod2lu_obj(lo)))) {
rc = lod_get_lmv_ea(env, lo);
- if (rc < sizeof(struct lmv_mds_md_v1))
+ if (rc < (typeof(rc))sizeof(struct lmv_mds_md_v1))
GOTO(out, rc = rc > 0 ? -EINVAL : rc);
buf->lb_buf = info->lti_ea_store;
lod->lod_sp_me = LUSTRE_SP_CLI;
/* Set up allocation policy (QoS and RR) */
- CFS_INIT_LIST_HEAD(&lod->lod_qos.lq_oss_list);
+ INIT_LIST_HEAD(&lod->lod_qos.lq_oss_list);
init_rwsem(&lod->lod_qos.lq_rw_sem);
lod->lod_qos.lq_dirty = 1;
lod->lod_qos.lq_rr.lqr_dirty = 1;
if (lod->lod_pools_hash_body == NULL)
RETURN(-ENOMEM);
- CFS_INIT_LIST_HEAD(&lod->lod_pool_list);
+ INIT_LIST_HEAD(&lod->lod_pool_list);
lod->lod_pool_count = 0;
rc = lod_ost_pool_init(&lod->lod_pool_info, 0);
if (rc)
int lod_pools_fini(struct lod_device *lod)
{
struct obd_device *obd = lod2obd(lod);
- cfs_list_t *pos, *tmp;
- struct pool_desc *pool;
+ struct pool_desc *pool, *tmp;
ENTRY;
- cfs_list_for_each_safe(pos, tmp, &lod->lod_pool_list) {
- pool = cfs_list_entry(pos, struct pool_desc, pool_list);
+ list_for_each_entry_safe(pool, tmp, &lod->lod_pool_list, pool_list) {
/* free pool structs */
CDEBUG(D_INFO, "delete pool %p\n", pool);
/* In the function below, .hs_keycmp resolves to