Whamcloud - gitweb
LU-1187 tests: Add DNE test cases in sanity.
[fs/lustre-release.git] / lustre / lov / lov_object.c
index 9b52f34..65e4466 100644 (file)
@@ -201,7 +201,7 @@ static int lov_init_raid0(const struct lu_env *env,
         if (r0->lo_sub != NULL) {
                 result = 0;
                 subconf->coc_inode = conf->coc_inode;
-                cfs_spin_lock_init(&r0->lo_sub_lock);
+               spin_lock_init(&r0->lo_sub_lock);
                 /*
                  * Create stripe cl_objects.
                  */
@@ -215,6 +215,9 @@ static int lov_init_raid0(const struct lu_env *env,
                         subdev = lovsub2cl_dev(dev->ld_target[ost_idx]);
                         subconf->u.coc_oinfo = oinfo;
                         LASSERTF(subdev != NULL, "not init ost %d\n", ost_idx);
+                       /* In the function below, .hs_keycmp resolves to
+                        * lu_obj_hop_keycmp() */
+                       /* coverity[overrun-buffer-val] */
                         stripe = lov_sub_find(env, subdev, ofid, subconf);
                         if (!IS_ERR(stripe))
                                 result = lov_init_sub(env, lov, stripe, r0, i);
@@ -230,6 +233,7 @@ static int lov_delete_empty(const struct lu_env *env, struct lov_object *lov,
                            union lov_layout_state *state)
 {
        LASSERT(lov->lo_type == LLT_EMPTY);
+       cl_object_prune(env, &lov->lo_cl);
        return 0;
 }
 
@@ -265,12 +269,12 @@ static void lov_subobject_kill(const struct lu_env *env, struct lov_object *lov,
                         /* this wait-queue is signaled at the end of
                          * lu_object_free(). */
                         cfs_set_current_state(CFS_TASK_UNINT);
-                        cfs_spin_lock(&r0->lo_sub_lock);
-                        if (r0->lo_sub[idx] == los) {
-                                cfs_spin_unlock(&r0->lo_sub_lock);
-                                cfs_waitq_wait(waiter, CFS_TASK_UNINT);
-                        } else {
-                                cfs_spin_unlock(&r0->lo_sub_lock);
+                       spin_lock(&r0->lo_sub_lock);
+                       if (r0->lo_sub[idx] == los) {
+                               spin_unlock(&r0->lo_sub_lock);
+                               cfs_waitq_wait(waiter, CFS_TASK_UNINT);
+                       } else {
+                               spin_unlock(&r0->lo_sub_lock);
                                 cfs_set_current_state(CFS_TASK_RUNNING);
                                 break;
                         }
@@ -297,12 +301,14 @@ static int lov_delete_raid0(const struct lu_env *env, struct lov_object *lov,
                 for (i = 0; i < r0->lo_nr; ++i) {
                         struct lovsub_object *los = r0->lo_sub[i];
 
-                        if (los != NULL)
+                        if (los != NULL) {
+                               cl_locks_prune(env, &los->lso_cl, 1);
                                 /*
                                  * If top-level object is to be evicted from
                                  * the cache, so are its sub-objects.
                                  */
                                 lov_subobject_kill(env, lov, los, i);
+                       }
                 }
         }
        RETURN(0);
@@ -385,8 +391,14 @@ static int lov_attr_get_raid0(const struct lu_env *env, struct cl_object *obj,
         ENTRY;
 
        /* this is called w/o holding type guard mutex, so it must be inside
-        * an on going IO otherwise lsm may be replaced. */
-       LASSERT(cfs_atomic_read(&lsm->lsm_refc) > 1);
+        * an on going IO otherwise lsm may be replaced.
+        * LU-2117: it turns out there exists one exception. For mmaped files,
+        * the lock of those files may be requested in the other file's IO
+        * context, and this function is called in ccc_lock_state(), it will
+        * hit this assertion.
+        * Anyway, it's still okay to call attr_get w/o type guard as layout
+        * can't go if locks exist. */
+       /* LASSERT(cfs_atomic_read(&lsm->lsm_refc) > 1); */
 
         if (!r0->lo_attr_valid) {
                 /*
@@ -430,7 +442,7 @@ const static struct lov_layout_operations lov_dispatch[] = {
                 .llo_install   = lov_install_empty,
                 .llo_print     = lov_print_empty,
                 .llo_page_init = lov_page_init_empty,
-                .llo_lock_init = NULL,
+                .llo_lock_init = lov_lock_init_empty,
                 .llo_io_init   = lov_io_init_empty,
                 .llo_getattr   = lov_attr_get_empty
         },
@@ -464,13 +476,13 @@ const static struct lov_layout_operations lov_dispatch[] = {
 static inline void lov_conf_freeze(struct lov_object *lov)
 {
        if (lov->lo_owner != cfs_current())
-               cfs_down_read(&lov->lo_type_guard);
+               down_read(&lov->lo_type_guard);
 }
 
 static inline void lov_conf_thaw(struct lov_object *lov)
 {
        if (lov->lo_owner != cfs_current())
-               cfs_up_read(&lov->lo_type_guard);
+               up_read(&lov->lo_type_guard);
 }
 
 #define LOV_2DISPATCH_MAYLOCK(obj, op, lock, ...)                       \
@@ -505,6 +517,20 @@ do {                                                                    \
        lov_conf_thaw(__obj);                                           \
 } while (0)
 
+static void lov_conf_lock(struct lov_object *lov)
+{
+       LASSERT(lov->lo_owner != cfs_current());
+       down_write(&lov->lo_type_guard);
+       LASSERT(lov->lo_owner == NULL);
+       lov->lo_owner = cfs_current();
+}
+
+static void lov_conf_unlock(struct lov_object *lov)
+{
+       lov->lo_owner = NULL;
+       up_write(&lov->lo_type_guard);
+}
+
 static int lov_layout_wait(const struct lu_env *env, struct lov_object *lov)
 {
        struct l_wait_info lwi = { 0 };
@@ -514,11 +540,17 @@ static int lov_layout_wait(const struct lu_env *env, struct lov_object *lov)
        if (!lov->lo_lsm_invalid || lsm == NULL)
                RETURN(0);
 
-       l_wait_event(lov->lo_waitq, cfs_atomic_read(&lsm->lsm_refc) == 1, &lwi);
+       LASSERT(cfs_atomic_read(&lsm->lsm_refc) > 0);
+       while (cfs_atomic_read(&lsm->lsm_refc) > 1) {
+               lov_conf_unlock(lov);
+               l_wait_event(lov->lo_waitq,
+                            cfs_atomic_read(&lsm->lsm_refc) == 1, &lwi);
+               lov_conf_lock(lov);
+       }
        RETURN(0);
 }
 
-static int lov_layout_change(const struct lu_env *env,
+static int lov_layout_change(const struct lu_env *unused,
                              struct lov_object *lov, enum lov_layout_type llt,
                              const struct cl_object_conf *conf)
 {
@@ -529,7 +561,7 @@ static int lov_layout_change(const struct lu_env *env,
 
        struct cl_object_header *hdr = cl_object_header(&lov->lo_cl);
        void *cookie;
-       struct lu_env *nested;
+       struct lu_env *env;
        int refcheck;
 
        LASSERT(0 <= lov->lo_type && lov->lo_type < ARRAY_SIZE(lov_dispatch));
@@ -537,13 +569,11 @@ static int lov_layout_change(const struct lu_env *env,
        ENTRY;
 
        cookie = cl_env_reenter();
-       nested = cl_env_get(&refcheck);
-       if (!IS_ERR(nested))
-               cl_object_prune(nested, &lov->lo_cl);
-       else
-               result = PTR_ERR(nested);
-       cl_env_put(nested, &refcheck);
-       cl_env_reexit(cookie);
+       env = cl_env_get(&refcheck);
+       if (IS_ERR(env)) {
+               cl_env_reexit(cookie);
+               RETURN(PTR_ERR(env));
+       }
 
        old_ops = &lov_dispatch[lov->lo_type];
        new_ops = &lov_dispatch[llt];
@@ -568,6 +598,9 @@ static int lov_layout_change(const struct lu_env *env,
                        /* this file becomes an EMPTY file. */
                }
        }
+
+       cl_env_put(env, &refcheck);
+       cl_env_reexit(cookie);
        RETURN(result);
 }
 
@@ -588,7 +621,7 @@ int lov_object_init(const struct lu_env *env, struct lu_object *obj,
         int result;
 
         ENTRY;
-        cfs_init_rwsem(&lov->lo_type_guard);
+       init_rwsem(&lov->lo_type_guard);
        cfs_waitq_init(&lov->lo_waitq);
 
         /* no locking is necessary, as object is being created */
@@ -603,32 +636,33 @@ int lov_object_init(const struct lu_env *env, struct lu_object *obj,
 static int lov_conf_set(const struct lu_env *env, struct cl_object *obj,
                         const struct cl_object_conf *conf)
 {
-       struct lov_stripe_md *lsm = conf->u.coc_md->lsm;
+       struct lov_stripe_md *lsm = NULL;
        struct lov_object *lov = cl2lov(obj);
        int result = 0;
        ENTRY;
 
-       /*
-        * Only LLT_EMPTY <-> LLT_RAID0 transitions are supported.
-        */
-       LASSERT(lov->lo_owner != cfs_current());
-       cfs_down_write(&lov->lo_type_guard);
-       LASSERT(lov->lo_owner == NULL);
-       lov->lo_owner = cfs_current();
-
+       lov_conf_lock(lov);
        if (conf->coc_invalidate) {
                lov->lo_lsm_invalid = 1;
                GOTO(out, result = 0);
        }
 
-       if (conf->coc_validate_only) {
-               if (!lov->lo_lsm_invalid)
-                       GOTO(out, result = 0);
+       if (conf->u.coc_md != NULL)
+               lsm = conf->u.coc_md->lsm;
 
-               lov_layout_wait(env, lov);
-               /* fall through to set up new layout */
+       if ((lsm == NULL && lov->lo_lsm == NULL) ||
+           (lsm != NULL && lov->lo_lsm != NULL &&
+            lov->lo_lsm->lsm_layout_gen == lsm->lsm_layout_gen)) {
+               lov->lo_lsm_invalid = 0;
+               GOTO(out, result = 0);
        }
 
+       /* will change layout */
+       lov_layout_wait(env, lov);
+
+       /*
+        * Only LLT_EMPTY <-> LLT_RAID0 transitions are supported.
+        */
        switch (lov->lo_type) {
        case LLT_EMPTY:
                if (lsm != NULL)
@@ -647,8 +681,7 @@ static int lov_conf_set(const struct lu_env *env, struct cl_object *obj,
        EXIT;
 
 out:
-       lov->lo_owner = NULL;
-       cfs_up_write(&lov->lo_type_guard);
+       lov_conf_unlock(lov);
        RETURN(result);
 }
 
@@ -681,8 +714,8 @@ static int lov_object_print(const struct lu_env *env, void *cookie,
 struct cl_page *lov_page_init(const struct lu_env *env, struct cl_object *obj,
                               struct cl_page *page, cfs_page_t *vmpage)
 {
-        return LOV_2DISPATCH(cl2lov(obj),
-                             llo_page_init, env, obj, page, vmpage);
+        return LOV_2DISPATCH_NOLOCK(cl2lov(obj),
+                                   llo_page_init, env, obj, page, vmpage);
 }
 
 /**
@@ -692,15 +725,9 @@ struct cl_page *lov_page_init(const struct lu_env *env, struct cl_object *obj,
 int lov_io_init(const struct lu_env *env, struct cl_object *obj,
                struct cl_io *io)
 {
-       struct lov_io *lio = lov_env_io(env);
-
        CL_IO_SLICE_CLEAN(lov_env_io(env), lis_cl);
-
-       /* hold lsm before initializing because io relies on it */
-       lio->lis_lsm = lov_lsm_addref(cl2lov(obj));
-
-       /* No need to lock because we've taken one refcount of layout.  */
-       return LOV_2DISPATCH_NOLOCK(cl2lov(obj), llo_io_init, env, obj, io);
+       return LOV_2DISPATCH_MAYLOCK(cl2lov(obj), llo_io_init,
+                                    !io->ci_ignore_layout, env, obj, io);
 }
 
 /**
@@ -781,10 +808,11 @@ struct lov_stripe_md *lov_lsm_addref(struct lov_object *lov)
        struct lov_stripe_md *lsm = NULL;
 
        lov_conf_freeze(lov);
-       if (!lov->lo_lsm_invalid && lov->lo_lsm != NULL) {
+       if (lov->lo_lsm != NULL) {
                lsm = lsm_addref(lov->lo_lsm);
-               CDEBUG(D_INODE, "lsm %p addref %d by %p.\n",
-                       lsm, cfs_atomic_read(&lsm->lsm_refc), cfs_current());
+               CDEBUG(D_INODE, "lsm %p addref %d/%d by %p.\n",
+                       lsm, cfs_atomic_read(&lsm->lsm_refc),
+                       lov->lo_lsm_invalid, cfs_current());
        }
        lov_conf_thaw(lov);
        return lsm;