void lfsck_lmv_put(const struct lu_env *env, struct lfsck_lmv *llmv)
{
if (llmv != NULL && atomic_dec_and_test(&llmv->ll_ref)) {
- if (llmv->ll_lslr != NULL)
+ if (llmv->ll_inline) {
+ struct lfsck_lmv_unit *llu;
+ struct lfsck_instance *lfsck;
+
+ llu = list_entry(llmv, struct lfsck_lmv_unit, llu_lmv);
+ lfsck = llu->llu_lfsck;
+
+ spin_lock(&lfsck->li_lock);
+ list_del(&llu->llu_link);
+ spin_unlock(&lfsck->li_lock);
+
+ lfsck_object_put(env, llu->llu_obj);
+
+ LASSERT(llmv->ll_lslr != NULL);
+
OBD_FREE_LARGE(llmv->ll_lslr,
- sizeof(struct lfsck_slave_lmv_rec) *
- llmv->ll_stripes_allocated);
+ sizeof(*llmv->ll_lslr) *
+ llmv->ll_stripes_allocated);
+ OBD_FREE_PTR(llu);
+ } else {
+ if (llmv->ll_lslr != NULL)
+ OBD_FREE_LARGE(llmv->ll_lslr,
+ sizeof(*llmv->ll_lslr) *
+ llmv->ll_stripes_allocated);
+
+ OBD_FREE_PTR(llmv);
+ }
+ }
+}
+
+/**
+ * Mark the specified directory as read-only by set LUSTRE_IMMUTABLE_FL.
+ *
+ * The caller has taken the ldlm lock on the @obj already.
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] obj pointer to the object to be handled
+ * \param[in] del_lmv true if need to drop the LMV EA
+ *
+ * \retval positive number if nothing to be done
+ * \retval zero for succeed
+ * \retval negative error number on failure
+ */
+static int lfsck_disable_master_lmv(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *obj, bool del_lmv)
+{
+ struct lfsck_thread_info *info = lfsck_env_info(env);
+ struct lu_attr *la = &info->lti_la;
+ struct lfsck_instance *lfsck = com->lc_lfsck;
+ struct dt_device *dev = lfsck_obj2dt_dev(obj);
+ struct thandle *th = NULL;
+ int rc = 0;
+ ENTRY;
- OBD_FREE_PTR(llmv);
+ th = dt_trans_create(env, dev);
+ if (IS_ERR(th))
+ GOTO(log, rc = PTR_ERR(th));
+
+ if (del_lmv) {
+ rc = dt_declare_xattr_del(env, obj, XATTR_NAME_LMV, th);
+ if (rc != 0)
+ GOTO(stop, rc);
}
+
+ la->la_valid = LA_FLAGS;
+ rc = dt_declare_attr_set(env, obj, la, th);
+ if (rc != 0)
+ GOTO(stop, rc);
+
+ rc = dt_trans_start_local(env, dev, th);
+ if (rc != 0)
+ GOTO(stop, rc);
+
+ dt_write_lock(env, obj, 0);
+ if (unlikely(lfsck_is_dead_obj(obj)))
+ GOTO(unlock, rc = 1);
+
+ if (lfsck->li_bookmark_ram.lb_param & LPF_DRYRUN)
+ GOTO(unlock, rc = 0);
+
+ if (del_lmv) {
+ rc = dt_xattr_del(env, obj, XATTR_NAME_LMV, th, BYPASS_CAPA);
+ if (rc != 0)
+ GOTO(unlock, rc);
+ }
+
+ rc = dt_attr_get(env, obj, la, BYPASS_CAPA);
+ if (rc == 0 && !(la->la_flags & LUSTRE_IMMUTABLE_FL)) {
+ la->la_valid = LA_FLAGS;
+ la->la_flags |= LUSTRE_IMMUTABLE_FL;
+ rc = dt_attr_set(env, obj, la, th, BYPASS_CAPA);
+ }
+
+ GOTO(unlock, rc);
+
+unlock:
+ dt_write_unlock(env, obj);
+
+stop:
+ dt_trans_stop(env, dev, th);
+
+log:
+ CDEBUG(D_LFSCK, "%s: namespace LFSCK set the master MDT-object of "
+ "the striped directory "DFID" as read-only: rc = %d\n",
+ lfsck_lfsck2name(lfsck), PFID(lfsck_dto2fid(obj)), rc);
+
+ if (rc <= 0) {
+ struct lfsck_namespace *ns = com->lc_file_ram;
+
+ ns->ln_flags |= LF_INCONSISTENT;
+ if (rc == 0)
+ ns->ln_striped_dirs_disabled++;
+ }
+
+ return rc;
}
static inline bool lfsck_is_valid_slave_lmv(struct lmv_mds_md_v1 *lmv)
lmv_is_known_hash_type(lmv->lmv_hash_type);
}
+/**
+ * Remove the striped directory's master LMV EA and mark it as read-only.
+ *
+ * Take ldlm lock on the striped directory before calling the
+ * lfsck_disable_master_lmv().
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] obj pointer to the striped directory to be handled
+ * \param[in] lnr pointer to the namespace request that contains the
+ * striped directory to be handled and other information
+ *
+ * \retval positive number if nothing to be done
+ * \retval zero for succeed
+ * \retval negative error number on failure
+ */
+static int lfsck_remove_lmv(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *obj,
+ struct lfsck_namespace_req *lnr)
+{
+ struct lustre_handle lh = { 0 };
+ int rc;
+
+ lnr->lnr_lmv->ll_ignore = 1;
+ rc = lfsck_ibits_lock(env, com->lc_lfsck, obj, &lh,
+ MDS_INODELOCK_UPDATE | MDS_INODELOCK_XATTR,
+ LCK_EX);
+ if (rc == 0) {
+ rc = lfsck_disable_master_lmv(env, com, obj, true);
+ lfsck_ibits_unlock(&lh, LCK_EX);
+ }
+
+ return rc;
+}
+
+/**
+ * Remove the name entry from the striped directory's master MDT-object.
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] dir pointer to the striped directory
+ * \param[in] fid the shard's FID which name entry will be removed
+ * \param[in] index the shard's index which name entry will be removed
+ *
+ * \retval positive number for repaired successfully
+ * \retval 0 if nothing to be repaired
+ * \retval negative error number on failure
+ */
+static int lfsck_remove_dirent(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *dir,
+ const struct lu_fid *fid, __u32 index)
+{
+ struct lfsck_thread_info *info = lfsck_env_info(env);
+ struct dt_object *obj;
+ int rc;
+
+ snprintf(info->lti_tmpbuf2, sizeof(info->lti_tmpbuf2), DFID":%u",
+ PFID(fid), index);
+ obj = lfsck_object_find_by_dev(env, com->lc_lfsck->li_bottom, fid);
+ if (IS_ERR(obj))
+ return PTR_ERR(obj);
+
+ rc = lfsck_namespace_repair_dirent(env, com, dir, obj,
+ info->lti_tmpbuf2, info->lti_tmpbuf2,
+ S_IFDIR, false, false);
+ lfsck_object_put(env, obj);
+ if (rc > 0) {
+ struct lfsck_namespace *ns = com->lc_file_ram;
+
+ ns->ln_dirent_repaired++;
+ }
+
+ return rc;
+}
+
+/**
+ * Remove old shard's name entry and refill the @lslr slot with new shard.
+ *
+ * Some old shard held the specified @lslr slot, but it is an invalid shard.
+ * This function will remove the bad shard's name entry, and refill the @lslr
+ * slot with the new shard.
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] dir pointer to the striped directory to be handled
+ * \param[in] lslr pointer to lfsck_disable_master_lmv slot which content
+ * will be replaced by the given information
+ * \param[in] lnr contain the shard's FID to be used to fill the
+ * @lslr slot, it also records the known max filled index
+ * and the known max stripe count
+ * \param[in] lmv contain the slave LMV EA to be used to fill the
+ * @lslr slot
+ * \param[in] index the old shard's index in the striped directory
+ * \param[in] flags the new shard's flags in the @lslr slot
+ *
+ * \retval zero for succeed
+ * \retval negative error number on failure
+ */
+static int lfsck_replace_lmv(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *dir,
+ struct lfsck_slave_lmv_rec *lslr,
+ struct lfsck_namespace_req *lnr,
+ struct lmv_mds_md_v1 *lmv,
+ __u32 index, __u32 flags)
+{
+ struct lfsck_lmv *llmv = lnr->lnr_lmv;
+ int rc;
+
+ rc = lfsck_remove_dirent(env, com, dir,
+ &lslr->lslr_fid, index);
+ if (rc < 0)
+ return rc;
+
+ lslr->lslr_fid = lnr->lnr_fid;
+ lslr->lslr_flags = flags;
+ lslr->lslr_stripe_count = lmv->lmv_stripe_count;
+ lslr->lslr_index = lmv->lmv_master_mdt_index;
+ lslr->lslr_hash_type = lmv->lmv_hash_type;
+ if (flags == LSLF_NONE) {
+ if (llmv->ll_hash_type == LMV_HASH_TYPE_UNKNOWN &&
+ lmv_is_known_hash_type(lmv->lmv_hash_type))
+ llmv->ll_hash_type = lmv->lmv_hash_type;
+
+ if (lslr->lslr_stripe_count <= LFSCK_LMV_MAX_STRIPES &&
+ llmv->ll_max_stripe_count < lslr->lslr_stripe_count)
+ llmv->ll_max_stripe_count = lslr->lslr_stripe_count;
+ }
+
+ return 0;
+}
+
+/**
+ * Record the slave LMV EA in the lfsck_lmv::ll_lslr.
+ *
+ * If the lfsck_lmv::ll_lslr slot corresponding to the given @shard_idx is free,
+ * then fill the slot with the given @lnr/@lmv/@flags directly (maybe need to
+ * extend the lfsck_lmv::ll_lslr buffer).
+ *
+ * If the lfsck_lmv::ll_lslr slot corresponding to the given @shard_idx is taken
+ * by other shard, then the LFSCK will try to resolve the conflict by checking
+ * the two conflict shards' flags, and try other possible slot (if one of them
+ * claims another possible @shard_idx).
+ *
+ * 1) If one of the two conflict shards can be recorded in another slot, then
+ * it is OK, go ahead. Otherwise,
+ *
+ * 2) If one of them is dangling name entry, then remove (one of) the dangling
+ * name entry (and replace related @lslr slot if needed). Otherwise,
+ *
+ * 3) If one of them has no slave LMV EA, then check whether the master LMV
+ * EA has ever been lost and re-generated (LMV_HASH_FLAG_LOST_LMV in the
+ * master LMV EA).
+ *
+ * 3.1) If yes, then it is possible that such object is not a real shard of
+ * the striped directory, instead, it was created by someone after the
+ * master LMV EA lost with the name that matches the shard naming rule.
+ * Then the LFSCK will remove the master LMV EA and mark the striped
+ * directory as read-only to allow those non-shard files to be visible
+ * to client.
+ *
+ * 3.2) If no, then remove (one of) the object what has no slave LMV EA.
+ *
+ * 4) If all above efforts cannot work, then the LFSCK cannot know how to
+ * recover the striped directory. To make the administrator can see the
+ * conflicts, the LFSCK will remove the master LMV EA and mark the striped
+ * directory as read-only.
+ *
+ * This function may be called recursively, to prevent overflow, we define
+ * LFSCK_REC_LMV_MAX_DEPTH to restrict the recursive call depth.
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] dir pointer to the striped directory to be handled
+ * \param[in] lnr contain the shard's FID to fill the @lslr slot,
+ * it also records the known max filled index and
+ * the known max stripe count
+ * \param[in] lmv pointer to the slave LMV EA to be recorded
+ * \param[in] shard_idx the shard's index used for locating the @lslr slot,
+ * it can be the index stored in the shard's name,
+ * it also can be the index stored in the slave LMV EA
+ * (for recursive case)
+ * \param[in] flags the shard's flags to be recorded in the @lslr slot
+ * to indicate the shard status, such as whether has
+ * slave LMV EA, whether dangling name entry, whether
+ * the name entry and slave LMV EA unmatched, and ect
+ * \param[in] flags2 when be called recursively, the @flags2 tells the
+ * former conflict shard's flags in the @lslr slot.
+ * \param[in,out] depth To prevent to be called recurisively too deep,
+ * we define the max depth can be called recursively
+ * (LFSCK_REC_LMV_MAX_DEPTH)
+ *
+ * \retval zero for succeed
+ * \retval "-ERANGE" for invalid @shard_idx
+ * \retval "-EEXIST" for the required lslr slot has been
+ * occupied by other shard
+ * \retval other negative error number on failure
+ */
+static int lfsck_record_lmv(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *dir,
+ struct lfsck_namespace_req *lnr,
+ struct lmv_mds_md_v1 *lmv, __u32 shard_idx,
+ __u32 flags, __u32 flags2, __u32 *depth)
+{
+ struct lfsck_instance *lfsck = com->lc_lfsck;
+ struct lfsck_lmv *llmv = lnr->lnr_lmv;
+ const struct lu_fid *fid = &lnr->lnr_fid;
+ struct lfsck_slave_lmv_rec *lslr;
+ struct lfsck_rec_lmv_save *lrls;
+ int index = shard_idx;
+ int rc = 0;
+ ENTRY;
+
+ CDEBUG(D_LFSCK, "%s: record slave LMV EA for the striped directory "
+ DFID": shard = "DFID", index = %u, flags = %u, flags2 = %u, "
+ "depth = %d\n", lfsck_lfsck2name(lfsck),
+ PFID(lfsck_dto2fid(dir)), PFID(fid),
+ index, flags, flags2, *depth);
+
+ if (index < 0 || index >= LFSCK_LMV_MAX_STRIPES)
+ RETURN(-ERANGE);
+
+ if (index >= llmv->ll_stripes_allocated) {
+ struct lfsck_slave_lmv_rec *new_lslr;
+ int new_stripes = index + 1;
+ size_t old_size = sizeof(*lslr) * llmv->ll_stripes_allocated;
+
+ OBD_ALLOC_LARGE(new_lslr, sizeof(*new_lslr) * new_stripes);
+ if (new_lslr == NULL) {
+ llmv->ll_failed = 1;
+
+ RETURN(-ENOMEM);
+ }
+
+ memcpy(new_lslr, llmv->ll_lslr, old_size);
+ OBD_FREE_LARGE(llmv->ll_lslr, old_size);
+ llmv->ll_stripes_allocated = new_stripes;
+ llmv->ll_lslr = new_lslr;
+ }
+
+ lslr = llmv->ll_lslr + index;
+ if (unlikely(lu_fid_eq(&lslr->lslr_fid, fid)))
+ RETURN(0);
+
+ if (fid_is_zero(&lslr->lslr_fid)) {
+ lslr->lslr_fid = *fid;
+ lslr->lslr_stripe_count = lmv->lmv_stripe_count;
+ lslr->lslr_index = lmv->lmv_master_mdt_index;
+ lslr->lslr_hash_type = lmv->lmv_hash_type;
+ lslr->lslr_flags = flags;
+ llmv->ll_stripes_filled++;
+ if (flags == LSLF_NONE) {
+ if (llmv->ll_hash_type == LMV_HASH_TYPE_UNKNOWN &&
+ lmv_is_known_hash_type(lmv->lmv_hash_type))
+ llmv->ll_hash_type = lmv->lmv_hash_type;
+
+ if (lslr->lslr_stripe_count <= LFSCK_LMV_MAX_STRIPES &&
+ llmv->ll_max_stripe_count < lslr->lslr_stripe_count)
+ llmv->ll_max_stripe_count =
+ lslr->lslr_stripe_count;
+ }
+
+ if (llmv->ll_max_filled_off < index)
+ llmv->ll_max_filled_off = index;
+
+ RETURN(0);
+ }
+
+ (*depth)++;
+ if (flags != LSLF_BAD_INDEX2)
+ LASSERTF(*depth == 1, "depth = %d\n", *depth);
+
+ /* Handle conflict cases. */
+ switch (lslr->lslr_flags) {
+ case LSLF_NONE:
+ case LSLF_BAD_INDEX2:
+ /* The existing one is a normal valid object. */
+ switch (flags) {
+ case LSLF_NONE:
+ /* The two 'valid' name entries claims the same
+ * index, the LFSCK cannot distinguish which one
+ * is correct. Then remove the master LMV EA to
+ * make all shards to be visible to client, and
+ * mark the master MDT-object as read-only. The
+ * administrator can handle the conflict with
+ * more human knowledge. */
+ rc = lfsck_remove_lmv(env, com, dir, lnr);
+ break;
+ case LSLF_BAD_INDEX2:
+ GOTO(out, rc = -EEXIST);
+ case LSLF_NO_LMVEA:
+
+no_lmvea:
+ if (llmv->ll_lmv.lmv_hash_type &
+ LMV_HASH_FLAG_LOST_LMV) {
+ /* If the master LMV EA was re-generated
+ * by the former LFSCK reparation, and
+ * before such reparation, someone has
+ * created the conflict object, but the
+ * LFSCK did not detect such conflict,
+ * then we have to remove the master
+ * LMV EA and mark the master MDT-object
+ * as read-only. The administrator can
+ * handle the conflict with more human
+ * knowledge. */
+ rc = lfsck_remove_lmv(env, com, dir, lnr);
+ } else {
+ /* Otherwise, remove the current name entry,
+ * and add its FID in the LFSCK tracing file
+ * for further processing. */
+ rc = lfsck_namespace_trace_update(env, com, fid,
+ LNTF_CHECK_PARENT, true);
+ if (rc == 0)
+ rc = lfsck_remove_dirent(env, com, dir,
+ fid, index);
+ }
+
+ break;
+ case LSLF_DANGLING:
+ /* Remove the current dangling name entry. */
+ rc = lfsck_remove_dirent(env, com, dir, fid, index);
+ break;
+ case LSLF_BAD_INDEX1:
+ index = lmv->lmv_master_mdt_index;
+ lmv->lmv_master_mdt_index = shard_idx;
+ /* The name entry claims an index that is conflict
+ * with a valid existing name entry, then try the
+ * index in the lmv recursively. */
+ rc = lfsck_record_lmv(env, com, dir, lnr, lmv, index,
+ LSLF_BAD_INDEX2, lslr->lslr_flags, depth);
+ lmv->lmv_master_mdt_index = index;
+ if (rc == -ERANGE || rc == -EEXIST)
+ /* The index in the lmv is invalid or
+ * also conflict with other. Then we do
+ * not know how to resolve the conflict.
+ * We will handle it as handle the case
+ * of 'LSLF_NONE' vs 'LSLF_NONE'. */
+ rc = lfsck_remove_lmv(env, com, dir, lnr);
+
+ break;
+ default:
+ break;
+ }
+
+ break;
+ case LSLF_NO_LMVEA:
+ /* The existing one has no slave LMV EA. */
+ switch (flags) {
+ case LSLF_NONE:
+
+none:
+ if (llmv->ll_lmv.lmv_hash_type &
+ LMV_HASH_FLAG_LOST_LMV) {
+ /* If the master LMV EA was re-generated
+ * by the former LFSCK reparation, and
+ * before such reparation, someone has
+ * created the conflict object, but the
+ * LFSCK did not detect such conflict,
+ * then we have to remove the master
+ * LMV EA and mark the master MDT-object
+ * as read-only. The administrator can
+ * handle the conflict with more human
+ * knowledge. */
+ rc = lfsck_remove_lmv(env, com, dir, lnr);
+ } else {
+ lrls = &lfsck->li_rec_lmv_save[*depth - 1];
+ lrls->lrls_fid = lslr->lslr_fid;
+ /* Otherwise, remove the existing name entry,
+ * and add its FID in the LFSCK tracing file
+ * for further processing. Refill the slot
+ * with current slave LMV EA. */
+ rc = lfsck_namespace_trace_update(env,
+ com, &lrls->lrls_fid,
+ LNTF_CHECK_PARENT, true);
+ if (rc == 0)
+ rc = lfsck_replace_lmv(env, com, dir,
+ lslr, lnr, lmv, index, flags);
+ }
+
+ break;
+ case LSLF_BAD_INDEX2:
+ if (flags2 >= lslr->lslr_flags)
+ GOTO(out, rc = -EEXIST);
+
+ goto none;
+ case LSLF_NO_LMVEA:
+ goto no_lmvea;
+ case LSLF_DANGLING:
+ /* Remove the current dangling name entry. */
+ rc = lfsck_remove_dirent(env, com, dir, fid, index);
+ break;
+ case LSLF_BAD_INDEX1:
+ index = lmv->lmv_master_mdt_index;
+ lmv->lmv_master_mdt_index = shard_idx;
+ /* The name entry claims an index that is conflict
+ * with a valid existing name entry, then try the
+ * index in the lmv recursively. */
+ rc = lfsck_record_lmv(env, com, dir, lnr, lmv, index,
+ LSLF_BAD_INDEX2, lslr->lslr_flags, depth);
+ lmv->lmv_master_mdt_index = index;
+ if (rc == -ERANGE || rc == -EEXIST) {
+ index = shard_idx;
+ goto no_lmvea;
+ }
+
+ break;
+ default:
+ break;
+ }
+
+ break;
+ case LSLF_DANGLING:
+ /* The existing one is a dangling name entry. */
+ switch (flags) {
+ case LSLF_NONE:
+ case LSLF_BAD_INDEX2:
+ case LSLF_NO_LMVEA:
+ /* Remove the existing dangling name entry.
+ * Refill the lslr slot with the given LMV. */
+ rc = lfsck_replace_lmv(env, com, dir, lslr, lnr,
+ lmv, index, flags);
+ break;
+ case LSLF_DANGLING:
+ /* Two dangling name entries conflict,
+ * remove the current one. */
+ rc = lfsck_remove_dirent(env, com, dir, fid, index);
+ break;
+ case LSLF_BAD_INDEX1:
+ index = lmv->lmv_master_mdt_index;
+ lmv->lmv_master_mdt_index = shard_idx;
+ /* The name entry claims an index that is conflict
+ * with a valid existing name entry, then try the
+ * index in the lmv recursively. */
+ rc = lfsck_record_lmv(env, com, dir, lnr, lmv, index,
+ LSLF_BAD_INDEX2, lslr->lslr_flags, depth);
+ lmv->lmv_master_mdt_index = index;
+ if (rc == -ERANGE || rc == -EEXIST)
+ /* If the index in the lmv is invalid or
+ * also conflict with other, then remove
+ * the existing dangling name entry.
+ * Refill the lslr slot with the given LMV. */
+ rc = lfsck_replace_lmv(env, com, dir, lslr, lnr,
+ lmv, shard_idx, flags);
+
+ break;
+ default:
+ break;
+ }
+
+ break;
+ case LSLF_BAD_INDEX1: {
+ if (*depth >= LFSCK_REC_LMV_MAX_DEPTH)
+ goto conflict;
+
+ lrls = &lfsck->li_rec_lmv_save[*depth - 1];
+ lrls->lrls_fid = lnr->lnr_fid;
+ lrls->lrls_lmv = *lmv;
+
+ lnr->lnr_fid = lslr->lslr_fid;
+ lmv->lmv_master_mdt_index = index;
+ lmv->lmv_stripe_count = lslr->lslr_stripe_count;
+ lmv->lmv_hash_type = lslr->lslr_hash_type;
+ index = lslr->lslr_index;
+
+ /* The existing one has another possible slot,
+ * try it recursively. */
+ rc = lfsck_record_lmv(env, com, dir, lnr, lmv, index,
+ LSLF_BAD_INDEX2, flags, depth);
+ *lmv = lrls->lrls_lmv;
+ lnr->lnr_fid = lrls->lrls_fid;
+ index = shard_idx;
+ if (rc != 0) {
+ if (rc == -ERANGE || rc == -EEXIST)
+ goto conflict;
+
+ break;
+ }
+
+ lslr->lslr_fid = *fid;
+ lslr->lslr_flags = flags;
+ lslr->lslr_stripe_count = lmv->lmv_stripe_count;
+ lslr->lslr_index = lmv->lmv_master_mdt_index;
+ lslr->lslr_hash_type = lmv->lmv_hash_type;
+ if (flags == LSLF_NONE) {
+ if (llmv->ll_hash_type == LMV_HASH_TYPE_UNKNOWN &&
+ lmv_is_known_hash_type(lmv->lmv_hash_type))
+ llmv->ll_hash_type = lmv->lmv_hash_type;
+
+ if (lslr->lslr_stripe_count <= LFSCK_LMV_MAX_STRIPES &&
+ llmv->ll_max_stripe_count < lslr->lslr_stripe_count)
+ llmv->ll_max_stripe_count =
+ lslr->lslr_stripe_count;
+ }
+
+ break;
+
+conflict:
+ switch (flags) {
+ case LSLF_NONE:
+ /* The two 'valid' name entries claims the same
+ * index, the LFSCK cannot distinguish which one
+ * is correct. Then remove the master LMV EA to
+ * make all shards to be visible to client, and
+ * mark the master MDT-object as read-only. The
+ * administrator can handle the conflict with
+ * more human knowledge. */
+ rc = lfsck_remove_lmv(env, com, dir, lnr);
+ break;
+ case LSLF_BAD_INDEX2:
+ GOTO(out, rc = -EEXIST);
+ case LSLF_NO_LMVEA:
+ goto no_lmvea;
+ case LSLF_DANGLING:
+ /* Remove the current dangling name entry. */
+ rc = lfsck_remove_dirent(env, com, dir, fid, index);
+ break;
+ case LSLF_BAD_INDEX1:
+ index = lmv->lmv_master_mdt_index;
+ lmv->lmv_master_mdt_index = shard_idx;
+ /* The name entry claims an index that is conflict
+ * with a valid existing name entry, then try the
+ * index in the lmv recursively. */
+ rc = lfsck_record_lmv(env, com, dir, lnr, lmv, index,
+ LSLF_BAD_INDEX2, lslr->lslr_flags, depth);
+ lmv->lmv_master_mdt_index = index;
+ if (rc == -ERANGE || rc == -EEXIST)
+ /* The index in the lmv is invalid or
+ * also conflict with other. Then we do
+ * not know how to resolve the conflict.
+ * We will handle it as handle the case
+ * of 'LSLF_NONE' vs 'LSLF_NONE'. */
+ rc = lfsck_remove_lmv(env, com, dir, lnr);
+
+ break;
+ }
+
+ break;
+ }
+ default:
+ break;
+ }
+
+ if (rc < 0)
+ llmv->ll_failed = 1;
+
+ GOTO(out, rc);
+
+out:
+ (*depth)--;
+
+ return rc > 0 ? 0 : rc;
+}
+
int lfsck_read_stripe_lmv(const struct lu_env *env, struct dt_object *obj,
struct lmv_mds_md_v1 *lmv)
{
}
/**
+ * Check whether allow to re-genereate the lost master LMV EA.
+ *
+ * If the master MDT-object of the striped directory lost its master LMV EA,
+ * then before the LFSCK repaired the striped directory, some ones may have
+ * created some objects (that are not normal shards of the striped directory)
+ * under the master MDT-object. If such case happend, then the LFSCK cannot
+ * re-generate the lost master LMV EA to keep those objects to be visible to
+ * client.
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] obj pointer to the master MDT-object to be checked
+ * \param[in] cfid the shard's FID used for verification
+ * \param[in] cidx the shard's index used for verification
+ *
+ * \retval positive number if not allow to re-generate LMV EA
+ * \retval zero if allow to re-generate LMV EA
+ * \retval negative error number on failure
+ */
+static int lfsck_allow_regenerate_master_lmv(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *obj,
+ const struct lu_fid *cfid,
+ __u32 cidx)
+{
+ struct lfsck_thread_info *info = lfsck_env_info(env);
+ struct lu_fid *tfid = &info->lti_fid3;
+ struct lfsck_instance *lfsck = com->lc_lfsck;
+ struct lu_dirent *ent =
+ (struct lu_dirent *)info->lti_key;
+ const struct dt_it_ops *iops;
+ struct dt_it *di;
+ __u64 cookie;
+ __u32 args;
+ int rc;
+ __u16 type;
+ ENTRY;
+
+ if (unlikely(!dt_try_as_dir(env, obj)))
+ RETURN(-ENOTDIR);
+
+ /* Check whether the shard and the master MDT-object matches or not. */
+ snprintf(info->lti_tmpbuf, sizeof(info->lti_tmpbuf), DFID":%u",
+ PFID(cfid), cidx);
+ rc = dt_lookup(env, obj, (struct dt_rec *)tfid,
+ (const struct dt_key *)info->lti_tmpbuf, BYPASS_CAPA);
+ if (rc != 0)
+ RETURN(rc);
+
+ if (!lu_fid_eq(tfid, cfid))
+ RETURN(-ENOENT);
+
+ args = lfsck->li_args_dir & ~(LUDA_VERIFY | LUDA_VERIFY_DRYRUN);
+ iops = &obj->do_index_ops->dio_it;
+ di = iops->init(env, obj, args, BYPASS_CAPA);
+ if (IS_ERR(di))
+ RETURN(PTR_ERR(di));
+
+ rc = iops->load(env, di, 0);
+ if (rc == 0)
+ rc = iops->next(env, di);
+ else if (rc > 0)
+ rc = 0;
+
+ if (rc != 0)
+ GOTO(out, rc);
+
+ do {
+ rc = iops->rec(env, di, (struct dt_rec *)ent, args);
+ if (rc == 0)
+ rc = lfsck_unpack_ent(ent, &cookie, &type);
+
+ if (rc != 0)
+ GOTO(out, rc);
+
+ /* skip dot and dotdot entries */
+ if (name_is_dot_or_dotdot(ent->lde_name, ent->lde_namelen))
+ goto next;
+
+ /* If the subdir name does not match the shard name rule, then
+ * it is quite possible that it is NOT a shard, but created by
+ * someone after the master MDT-object lost the master LMV EA.
+ * But it is also possible that the subdir name entry crashed,
+ * under such double failure cases, the LFSCK cannot know how
+ * to repair the inconsistency. For data safe, the LFSCK will
+ * mark the master MDT-object as read-only. The administrator
+ * can fix the bad shard name manually, then run LFSCK again.
+ *
+ * XXX: If the subdir name matches the shard name rule, but it
+ * is not a real shard of the striped directory, instead,
+ * it was created by someone after the master MDT-object
+ * lost the LMV EA, then re-generating the master LMV EA
+ * will cause such subdir to be invisible to client, and
+ * if its index occupies some lost shard index, then the
+ * LFSCK will use it to replace the bad shard, and cause
+ * the subdir (itself) to be invisible for ever. */
+ if (lfsck_shard_name_to_index(env, ent->lde_name,
+ ent->lde_namelen, type, &ent->lde_fid) < 0)
+ GOTO(out, rc = 1);
+
+next:
+ rc = iops->next(env, di);
+ } while (rc == 0);
+
+ GOTO(out, rc = 0);
+
+out:
+ iops->put(env, di);
+ iops->fini(env, di);
+
+ return rc;
+}
+
+/**
+ * Notify remote LFSCK instance that the object's LMV EA has been updated.
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] obj pointer to the object on which the LMV EA will be set
+ * \param[in] event indicate either master or slave LMV EA has been updated
+ * \param[in] flags indicate which element(s) in the LMV EA has been updated
+ * \param[in] index the MDT index on which the LFSCK instance to be notified
+ *
+ * \retval positive number if nothing to be done
+ * \retval zero for succeed
+ * \retval negative error number on failure
+ */
+static int lfsck_namespace_notify_lmv_remote(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *obj,
+ __u32 event, __u32 flags,
+ __u32 index)
+{
+ struct lfsck_request *lr = &lfsck_env_info(env)->lti_lr;
+ const struct lu_fid *fid = lfsck_dto2fid(obj);
+ struct lfsck_instance *lfsck = com->lc_lfsck;
+ struct lfsck_tgt_desc *ltd = NULL;
+ struct ptlrpc_request *req = NULL;
+ int rc;
+ ENTRY;
+
+ ltd = lfsck_tgt_get(&lfsck->li_mdt_descs, index);
+ if (ltd == NULL)
+ GOTO(out, rc = -ENODEV);
+
+ req = ptlrpc_request_alloc(class_exp2cliimp(ltd->ltd_exp),
+ &RQF_LFSCK_NOTIFY);
+ if (req == NULL)
+ GOTO(out, rc = -ENOMEM);
+
+ rc = ptlrpc_request_pack(req, LUSTRE_OBD_VERSION, LFSCK_NOTIFY);
+ if (rc != 0) {
+ ptlrpc_request_free(req);
+
+ GOTO(out, rc);
+ }
+
+ lr = req_capsule_client_get(&req->rq_pill, &RMF_LFSCK_REQUEST);
+ memset(lr, 0, sizeof(*lr));
+ lr->lr_event = event;
+ lr->lr_index = lfsck_dev_idx(lfsck->li_bottom);
+ lr->lr_active = LFSCK_TYPE_NAMESPACE;
+ lr->lr_fid = *fid;
+ lr->lr_flags = flags;
+
+ ptlrpc_request_set_replen(req);
+ rc = ptlrpc_queue_wait(req);
+ ptlrpc_req_finished(req);
+
+ GOTO(out, rc = (rc == -ENOENT ? 1 : rc));
+
+out:
+ CDEBUG(D_LFSCK, "%s: namespace LFSCK notify LMV EA updated for the "
+ "object "DFID" on MDT %x remotely with event %u, flags %u: "
+ "rc = %d\n", lfsck_lfsck2name(lfsck), PFID(fid), index,
+ event, flags, rc);
+
+ if (ltd != NULL)
+ lfsck_tgt_put(ltd);
+
+ return rc;
+}
+
+/**
+ * Generate request for local LFSCK instance to rescan the striped directory.
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] obj pointer to the striped directory to be rescanned
+ *
+ * \retval positive number if nothing to be done
+ * \retval zero for succeed
+ * \retval negative error number on failure
+ */
+int lfsck_namespace_notify_lmv_master_local(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *obj)
+{
+ struct lfsck_instance *lfsck = com->lc_lfsck;
+ struct lfsck_namespace *ns = com->lc_file_ram;
+ struct lmv_mds_md_v1 *lmv4 = &lfsck_env_info(env)->lti_lmv4;
+ struct lfsck_lmv_unit *llu;
+ struct lfsck_lmv *llmv;
+ struct lfsck_slave_lmv_rec *lslr;
+ int count = 0;
+ int rc;
+ ENTRY;
+
+ if (lfsck->li_bookmark_ram.lb_param & LPF_DRYRUN)
+ RETURN(0);
+
+ rc = lfsck_read_stripe_lmv(env, obj, lmv4);
+ if (rc != 0)
+ RETURN(rc);
+
+ OBD_ALLOC_PTR(llu);
+ if (unlikely(llu == NULL))
+ RETURN(-ENOMEM);
+
+ if (lmv4->lmv_stripe_count < 1)
+ count = LFSCK_LMV_DEF_STRIPES;
+ else if (lmv4->lmv_stripe_count > LFSCK_LMV_MAX_STRIPES)
+ count = LFSCK_LMV_MAX_STRIPES;
+ else
+ count = lmv4->lmv_stripe_count;
+
+ OBD_ALLOC_LARGE(lslr, sizeof(struct lfsck_slave_lmv_rec) * count);
+ if (lslr == NULL) {
+ OBD_FREE_PTR(llu);
+
+ RETURN(-ENOMEM);
+ }
+
+ INIT_LIST_HEAD(&llu->llu_link);
+ llu->llu_lfsck = lfsck;
+ llu->llu_obj = lfsck_object_get(obj);
+ llmv = &llu->llu_lmv;
+ llmv->ll_lmv_master = 1;
+ llmv->ll_inline = 1;
+ atomic_set(&llmv->ll_ref, 1);
+ llmv->ll_stripes_allocated = count;
+ llmv->ll_hash_type = LMV_HASH_TYPE_UNKNOWN;
+ llmv->ll_lslr = lslr;
+ llmv->ll_lmv = *lmv4;
+
+ down_write(&com->lc_sem);
+ if (ns->ln_status != LS_SCANNING_PHASE1 &&
+ ns->ln_status != LS_SCANNING_PHASE2) {
+ ns->ln_striped_dirs_skipped++;
+ up_write(&com->lc_sem);
+ lfsck_lmv_put(env, llmv);
+ } else {
+ ns->ln_striped_dirs_repaired++;
+ spin_lock(&lfsck->li_lock);
+ list_add_tail(&llu->llu_link, &lfsck->li_list_lmv);
+ spin_unlock(&lfsck->li_lock);
+ up_write(&com->lc_sem);
+ }
+
+ RETURN(0);
+}
+
+/**
* Set master LMV EA for the specified striped directory.
*
+ * First, if the master MDT-object of a striped directory lost its LMV EA,
+ * then there may be some users have created some files under the master
+ * MDT-object directly. Under such case, the LFSCK cannot re-generate LMV
+ * EA for the master MDT-object, because we should keep the existing files
+ * to be visible to client. Then the LFSCK will mark the striped directory
+ * as read-only and keep it there to be handled by administrator manually.
+ *
+ * If nobody has created files under the master MDT-object of the striped
+ * directory, then we will set the master LMV EA and generate a new rescan
+ * (the striped directory) request that will be handled later by the LFSCK
+ * instance on the MDT later.
+ *
* \param[in] env pointer to the thread context
* \param[in] com pointer to the lfsck component
* \param[in] dir pointer to the object on which the LMV EA will be set
{
struct lfsck_thread_info *info = lfsck_env_info(env);
struct lmv_mds_md_v1 *lmv3 = &info->lti_lmv3;
+ struct lu_seq_range *range = &info->lti_range;
struct lfsck_instance *lfsck = com->lc_lfsck;
+ struct seq_server_site *ss =
+ lu_site2seq(lfsck->li_bottom->dd_lu_dev.ld_site);
struct dt_object *obj;
struct lustre_handle lh = { 0 };
int pidx = -1;
if (unlikely(obj == NULL))
RETURN(-ENOENT);
- if (dt_object_remote(obj)) {
- struct lu_seq_range *range = &info->lti_range;
- struct seq_server_site *ss =
- lu_site2seq(lfsck->li_bottom->dd_lu_dev.ld_site);
-
- fld_range_set_mdt(range);
- rc = fld_server_lookup(env, ss->ss_server_fld,
- fid_seq(lfsck_dto2fid(obj)), range);
- if (rc != 0)
- GOTO(log, rc);
-
- pidx = range->lsr_index;
- } else {
- pidx = lfsck_dev_idx(lfsck->li_bottom);
- }
-
- /* XXX: it will be improved with subsequent patches landed. */
+ fld_range_set_mdt(range);
+ rc = fld_server_lookup(env, ss->ss_server_fld,
+ fid_seq(lfsck_dto2fid(obj)), range);
+ if (rc != 0)
+ GOTO(log, rc);
+ pidx = range->lsr_index;
rc = lfsck_ibits_lock(env, lfsck, obj, &lh,
MDS_INODELOCK_UPDATE | MDS_INODELOCK_XATTR,
LCK_EX);
GOTO(log, rc);
rc = lfsck_read_stripe_lmv(env, obj, lmv3);
- if (rc != 0)
+ if (rc == -ENODATA) {
+ if (!(flags & LEF_SET_LMV_ALL))
+ GOTO(log, rc);
+
+ *lmv3 = *lmv;
+ } else if (rc == 0) {
+ if (flags & LEF_SET_LMV_ALL)
+ GOTO(log, rc = 1);
+
+ if (flags & LEF_SET_LMV_HASH)
+ lmv3->lmv_hash_type = lmv->lmv_hash_type;
+ } else {
GOTO(log, rc);
+ }
- lmv3->lmv_hash_type = lmv->lmv_hash_type;
lmv3->lmv_magic = LMV_MAGIC;
lmv3->lmv_master_mdt_index = pidx;
+ if (flags & LEF_SET_LMV_ALL) {
+ rc = lfsck_allow_regenerate_master_lmv(env, com, obj,
+ cfid, cidx);
+ if (rc > 0) {
+ rc = lfsck_disable_master_lmv(env, com, obj, false);
+
+ GOTO(log, rc = (rc == 0 ? 1 : rc));
+ }
+
+ if (rc < 0)
+ GOTO(log, rc);
+
+ /* To indicate that the master has ever lost LMV EA. */
+ lmv3->lmv_hash_type |= LMV_HASH_FLAG_LOST_LMV;
+ }
+
rc = lfsck_namespace_update_lmv(env, com, obj, lmv3, true);
+ if (rc == 0 && flags & LEF_SET_LMV_ALL) {
+ if (dt_object_remote(obj))
+ rc = lfsck_namespace_notify_lmv_remote(env, com, obj,
+ LE_SET_LMV_MASTER, 0, pidx);
+ else
+ rc = lfsck_namespace_notify_lmv_master_local(env, com,
+ obj);
+ }
GOTO(log, rc);
}
/**
+ * Scan the shard of a striped directory for name hash verification.
+ *
+ * During the first-stage scanning, if the LFSCK cannot make sure whether
+ * the shard of a stripe directory contains valid slave LMV EA or not, then
+ * it will skip the name hash verification for this shard temporarily, and
+ * record the shard's FID in the LFSCK tracing file. As the LFSCK processing,
+ * the slave LMV EA may has been verified/fixed by LFSCK instance on master.
+ * Then in the second-stage scanning, the shard will be re-scanned, and for
+ * every name entry under the shard, the name hash will be verified, and for
+ * unmatched name entry, the LFSCK will try to fix it.
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] child pointer to the directory object to be handled
+ *
+ * \retval positive number for scanning successfully
+ * \retval zero for the scanning is paused
+ * \retval negative error number on failure
+ */
+int lfsck_namespace_scan_shard(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *child)
+{
+ struct lfsck_thread_info *info = lfsck_env_info(env);
+ struct lmv_mds_md_v1 *lmv = &info->lti_lmv;
+ struct lfsck_instance *lfsck = com->lc_lfsck;
+ struct lfsck_namespace *ns = com->lc_file_ram;
+ struct ptlrpc_thread *thread = &lfsck->li_thread;
+ struct lu_dirent *ent =
+ (struct lu_dirent *)info->lti_key;
+ struct lfsck_bookmark *bk = &lfsck->li_bookmark_ram;
+ struct lfsck_lmv *llmv = NULL;
+ const struct dt_it_ops *iops;
+ struct dt_it *di;
+ __u64 cookie;
+ __u32 args;
+ int rc;
+ __u16 type;
+ ENTRY;
+
+ rc = lfsck_read_stripe_lmv(env, child, lmv);
+ if (rc != 0)
+ RETURN(rc == -ENODATA ? 1 : rc);
+
+ if (lmv->lmv_magic != LMV_MAGIC_STRIPE)
+ RETURN(1);
+
+ if (unlikely(!dt_try_as_dir(env, child)))
+ RETURN(-ENOTDIR);
+
+ OBD_ALLOC_PTR(llmv);
+ if (llmv == NULL)
+ RETURN(-ENOMEM);
+
+ llmv->ll_lmv_slave = 1;
+ llmv->ll_lmv_verified = 1;
+ llmv->ll_lmv = *lmv;
+ atomic_set(&llmv->ll_ref, 1);
+
+ args = lfsck->li_args_dir & ~(LUDA_VERIFY | LUDA_VERIFY_DRYRUN);
+ iops = &child->do_index_ops->dio_it;
+ di = iops->init(env, child, args, BYPASS_CAPA);
+ if (IS_ERR(di))
+ GOTO(out, rc = PTR_ERR(di));
+
+ rc = iops->load(env, di, 0);
+ if (rc == 0)
+ rc = iops->next(env, di);
+ else if (rc > 0)
+ rc = 0;
+
+ while (rc == 0) {
+ if (CFS_FAIL_TIMEOUT(OBD_FAIL_LFSCK_DELAY3, cfs_fail_val) &&
+ unlikely(!thread_is_running(thread)))
+ GOTO(out, rc = 0);
+
+ rc = iops->rec(env, di, (struct dt_rec *)ent, args);
+ if (rc == 0)
+ rc = lfsck_unpack_ent(ent, &cookie, &type);
+
+ if (rc != 0) {
+ if (bk->lb_param & LPF_FAILOUT)
+ GOTO(out, rc);
+
+ goto next;
+ }
+
+ /* skip dot and dotdot entries */
+ if (name_is_dot_or_dotdot(ent->lde_name, ent->lde_namelen))
+ goto next;
+
+ if (!lfsck_is_valid_slave_name_entry(env, llmv, ent->lde_name,
+ ent->lde_namelen)) {
+ ns->ln_flags |= LF_INCONSISTENT;
+ rc = lfsck_namespace_repair_bad_name_hash(env, com,
+ child, llmv, ent->lde_name);
+ if (rc >= 0)
+ ns->ln_name_hash_repaired++;
+ }
+
+ if (rc < 0 && bk->lb_param & LPF_FAILOUT)
+ GOTO(out, rc);
+
+ /* Rate control. */
+ lfsck_control_speed(lfsck);
+ if (unlikely(!thread_is_running(thread)))
+ GOTO(out, rc = 0);
+
+ if (OBD_FAIL_CHECK(OBD_FAIL_LFSCK_FATAL2)) {
+ spin_lock(&lfsck->li_lock);
+ thread_set_flags(thread, SVC_STOPPING);
+ spin_unlock(&lfsck->li_lock);
+
+ GOTO(out, rc = -EINVAL);
+ }
+
+next:
+ rc = iops->next(env, di);
+ }
+
+ GOTO(out, rc);
+
+out:
+ iops->put(env, di);
+ iops->fini(env, di);
+ lfsck_lmv_put(env, llmv);
+
+ return rc;
+}
+
+/**
* Verify the slave object's (of striped directory) LMV EA.
*
* For the slave object of a striped directory, before traversing the shard
char *name = info->lti_key;
char *name2;
struct lu_fid *pfid = &info->lti_fid3;
- struct lu_fid *tfid = &info->lti_fid4;
const struct lu_fid *cfid = lfsck_dto2fid(obj);
+ struct lu_fid tfid;
struct lfsck_instance *lfsck = com->lc_lfsck;
struct lmv_mds_md_v1 *clmv = &llmv->ll_lmv;
struct lmv_mds_md_v1 *plmv = &info->lti_lmv;
GOTO(out, rc);
}
+ if (unlikely(!dt_try_as_dir(env, parent)))
+ GOTO(out, rc = -ENOTDIR);
+
rc = lfsck_read_stripe_lmv(env, parent, plmv);
if (rc != 0) {
int rc1;
/* If the parent has no LMV EA, then it maybe because:
* 1) The parent lost the LMV EA.
* 2) The child claims a wrong (slave) LMV EA. */
-
- /* XXX: to be improved. */
- rc = 0;
+ if (rc == -ENODATA)
+ rc = lfsck_namespace_set_lmv_master(env, com, parent,
+ clmv, cfid, clmv->lmv_master_mdt_index,
+ LEF_SET_LMV_ALL);
+ else
+ rc = 0;
rc1 = lfsck_namespace_trace_update(env, com, cfid,
LNTF_UNCERTAIN_LMV, true);
PFID(cfid), clmv->lmv_master_mdt_index);
name2 = info->lti_tmpbuf2;
- rc = lfsck_links_get_first(env, obj, name, tfid);
- if (rc == 0 && strcmp(name, name2) == 0 && lu_fid_eq(pfid, tfid)) {
+ rc = lfsck_links_get_first(env, obj, name, &tfid);
+ if (rc == 0 && strcmp(name, name2) == 0 && lu_fid_eq(pfid, &tfid)) {
llmv->ll_lmv_verified = 1;
GOTO(out, rc);
}
- rc = dt_lookup(env, parent, (struct dt_rec *)tfid,
+ rc = dt_lookup(env, parent, (struct dt_rec *)&tfid,
(const struct dt_key *)name2, BYPASS_CAPA);
- if (rc != 0 || !lu_fid_eq(cfid, tfid))
+ if (rc != 0 || !lu_fid_eq(cfid, &tfid))
rc = lfsck_namespace_trace_update(env, com, cfid,
LNTF_UNCERTAIN_LMV, true);
else
/**
* Double scan the striped directory or the shard.
*
+ * All the shards' under the given striped directory or its shard have
+ * been scanned, the LFSCK has got the global knownledge about the LMV
+ * EA consistency.
+ *
+ * If the target is one shard of a striped directory, then only needs to
+ * update related tracing file.
+ *
+ * If the target is the master MDT-object of a striped directory, then the
+ * LFSCK will make the decision about whether the master LMV EA is invalid
+ * or not, and repair it if inconsistenct; for every shard of the striped
+ * directory, whether the slave LMV EA is invalid or not, and repair it if
+ * inconsistent.
+ *
* \param[in] env pointer to the thread context
* \param[in] com pointer to the lfsck component
+ * \paran[in] dir pointer to the striped directory or its shard to be
+ * rescanned
* \param[in] lnr pointer to the namespace request that contains the
* striped directory or the shard
*
*/
int lfsck_namespace_striped_dir_rescan(const struct lu_env *env,
struct lfsck_component *com,
+ struct dt_object *dir,
struct lfsck_namespace_req *lnr)
{
+ struct lfsck_thread_info *info = lfsck_env_info(env);
+ struct lfsck_instance *lfsck = com->lc_lfsck;
struct lfsck_namespace *ns = com->lc_file_ram;
struct lfsck_lmv *llmv = lnr->lnr_lmv;
- struct dt_object *dir = lnr->lnr_obj;
+ struct lmv_mds_md_v1 *lmv = &llmv->ll_lmv;
+ struct lmv_mds_md_v1 *lmv2 = &info->lti_lmv2;
+ const struct lu_fid *pfid = lfsck_dto2fid(dir);
+ struct lu_seq_range *range = &info->lti_range;
+ struct seq_server_site *ss =
+ lu_site2seq(lfsck->li_bottom->dd_lu_dev.ld_site);
+ __u32 stripe_count;
+ __u32 hash_type;
+ int rc = 0;
+ int i;
ENTRY;
- /* XXX: it will be improved with subsequent patches landed. */
-
- if (llmv->ll_lmv_slave && llmv->ll_lmv_verified) {
- ns->ln_striped_shards_scanned++;
- lfsck_namespace_trace_update(env, com,
+ if (llmv->ll_lmv_slave) {
+ if (llmv->ll_lmv_verified) {
+ ns->ln_striped_shards_scanned++;
+ lfsck_namespace_trace_update(env, com,
lfsck_dto2fid(dir),
LNTF_UNCERTAIN_LMV |
LNTF_RECHECK_NAME_HASH, false);
+ }
+
+ RETURN(0);
}
- RETURN(0);
+ /* Either the striped directory has been disabled or only part of
+ * the striped directory have been scanned. The LFSCK cannot repair
+ * something based on incompleted knowledge. So skip it. */
+ if (llmv->ll_ignore || llmv->ll_exit_value <= 0)
+ RETURN(0);
+
+ /* There ever been some failure, as to the LFSCK cannot know whether
+ * it has got the global knowledge about the LMV EA consistency or not,
+ * so it cannot make reparation about the incompleted knowledge. */
+ if (llmv->ll_failed) {
+ ns->ln_striped_dirs_scanned++;
+ ns->ln_striped_dirs_failed++;
+
+ RETURN(0);
+ }
+
+ if (lmv->lmv_stripe_count > LFSCK_LMV_MAX_STRIPES)
+ stripe_count = max(llmv->ll_max_filled_off + 1,
+ llmv->ll_max_stripe_count);
+ else
+ stripe_count = max(llmv->ll_max_filled_off + 1,
+ lmv->lmv_stripe_count);
+
+ if (lmv->lmv_stripe_count != stripe_count) {
+ lmv->lmv_stripe_count = stripe_count;
+ llmv->ll_lmv_updated = 1;
+ }
+
+ if (!lmv_is_known_hash_type(lmv->lmv_hash_type) &&
+ !(lmv->lmv_hash_type & LMV_HASH_FLAG_BAD_TYPE) &&
+ lmv_is_known_hash_type(llmv->ll_hash_type)) {
+ hash_type = llmv->ll_hash_type & LMV_HASH_TYPE_MASK;
+ lmv->lmv_hash_type = llmv->ll_hash_type;
+ llmv->ll_lmv_updated = 1;
+ } else {
+ hash_type = lmv->lmv_hash_type & LMV_HASH_TYPE_MASK;
+ if (!lmv_is_known_hash_type(hash_type))
+ hash_type = LMV_HASH_TYPE_UNKNOWN;
+ }
+
+ if (llmv->ll_lmv_updated) {
+ lmv->lmv_layout_version++;
+ rc = lfsck_namespace_update_lmv(env, com, dir, lmv, false);
+ if (rc != 0)
+ RETURN(rc);
+
+ ns->ln_striped_dirs_scanned++;
+ ns->ln_striped_dirs_repaired++;
+ }
+
+ fld_range_set_mdt(range);
+ for (i = 0; i <= llmv->ll_max_filled_off; i++) {
+ struct dt_object *obj = NULL;
+ struct lfsck_slave_lmv_rec *lslr = llmv->ll_lslr + i;
+ const struct lu_fid *cfid = &lslr->lslr_fid;
+ const struct lu_name *cname;
+ struct linkea_data ldata = { NULL };
+ int len;
+ int rc1 = 0;
+ bool repair_linkea = false;
+ bool repair_lmvea = false;
+ bool rename = false;
+ bool create = false;
+ bool linkea_repaired = false;
+ bool lmvea_repaired = false;
+ bool rename_repaired = false;
+ bool create_repaired = false;
+
+ /* LMV EA hole. */
+ if (fid_is_zero(cfid))
+ continue;
+
+ len = snprintf(info->lti_tmpbuf, sizeof(info->lti_tmpbuf),
+ DFID":%u", PFID(cfid), i);
+ cname = lfsck_name_get_const(env, info->lti_tmpbuf, len);
+ memcpy(lnr->lnr_name, info->lti_tmpbuf, len);
+
+ obj = lfsck_object_find_bottom_nowait(env, lfsck, cfid);
+ if (IS_ERR(obj)) {
+ if (lfsck_is_dead_obj(dir))
+ RETURN(0);
+
+ rc1 = PTR_ERR(obj);
+ goto next;
+ }
+
+ switch (lslr->lslr_flags) {
+ case LSLF_NONE:
+ if (llmv->ll_inline ||
+ lslr->lslr_stripe_count != stripe_count ||
+ (lslr->lslr_hash_type & LMV_HASH_TYPE_MASK) !=
+ hash_type)
+ repair_lmvea = true;
+ break;
+ case LSLF_BAD_INDEX2:
+ /* The index in the slave LMV EA is right,
+ * the name entry should be updated. */
+ rename = true;
+ snprintf(info->lti_tmpbuf2, sizeof(info->lti_tmpbuf2),
+ DFID":%u", PFID(cfid), lslr->lslr_index);
+ if (llmv->ll_inline ||
+ lslr->lslr_stripe_count != stripe_count ||
+ (lslr->lslr_hash_type & LMV_HASH_TYPE_MASK) !=
+ hash_type)
+ repair_lmvea = true;
+ break;
+ case LSLF_BAD_INDEX1:
+ /* The index in the name entry is right,
+ * the slave LMV EA should be updated. */
+ case LSLF_NO_LMVEA:
+ repair_lmvea = true;
+ break;
+ case LSLF_DANGLING:
+ create = true;
+ goto repair;
+ default:
+ break;
+ }
+
+ rc1 = lfsck_links_read(env, obj, &ldata);
+ if (rc1 == -ENOENT) {
+ create = true;
+ goto repair;
+ }
+
+ if (rc1 == -EINVAL || rc1 == -ENODATA) {
+ repair_linkea = true;
+ goto repair;
+ }
+
+ if (rc1 != 0)
+ goto next;
+
+ if (ldata.ld_leh->leh_reccount != 1) {
+ repair_linkea = true;
+ goto repair;
+ }
+
+ rc1 = linkea_links_find(&ldata, cname, pfid);
+ if (rc1 != 0)
+ repair_linkea = true;
+
+repair:
+ if (create) {
+ rc1 = lfsck_namespace_repair_dangling(env, com, dir,
+ obj, lnr);
+ if (rc1 >= 0) {
+ create_repaired = true;
+ if (rc == 0)
+ ns->ln_dangling_repaired++;
+ }
+ }
+
+ if (repair_lmvea) {
+ *lmv2 = *lmv;
+ lmv2->lmv_magic = LMV_MAGIC_STRIPE;
+ lmv2->lmv_stripe_count = stripe_count;
+ lmv2->lmv_master_mdt_index = i;
+ lmv2->lmv_hash_type = hash_type;
+
+ rc1 = lfsck_namespace_update_lmv(env, com, obj,
+ lmv2, false);
+ if (rc1 < 0)
+ goto next;
+
+ if (dt_object_remote(obj)) {
+ rc1 = fld_server_lookup(env, ss->ss_server_fld,
+ fid_seq(lfsck_dto2fid(obj)), range);
+ if (rc1 != 0)
+ goto next;
+
+ rc1 = lfsck_namespace_notify_lmv_remote(env,
+ com, obj, LE_SET_LMV_SLAVE, 0,
+ range->lsr_index);
+ } else {
+ ns->ln_striped_shards_repaired++;
+ rc1 = lfsck_namespace_trace_update(env, com,
+ cfid, LNTF_RECHECK_NAME_HASH, true);
+ }
+
+ if (rc1 < 0)
+ goto next;
+
+ if (rc1 >= 0)
+ lmvea_repaired = true;
+ } else if (llmv->ll_inline) {
+ if (dt_object_remote(obj)) {
+ rc1 = fld_server_lookup(env, ss->ss_server_fld,
+ fid_seq(lfsck_dto2fid(obj)), range);
+ if (rc1 != 0)
+ goto next;
+
+ /* The slave LMV EA on the remote shard is
+ * correct, just notify the LFSCK instance
+ * on such MDT to re-verify the name_hash. */
+ rc1 = lfsck_namespace_notify_lmv_remote(env,
+ com, obj, LE_SET_LMV_SLAVE,
+ LEF_RECHECK_NAME_HASH,
+ range->lsr_index);
+ } else {
+ rc1 = lfsck_namespace_trace_update(env, com,
+ cfid, LNTF_RECHECK_NAME_HASH, true);
+ }
+
+ if (rc1 < 0)
+ goto next;
+ }
+
+ if (rename) {
+ rc1 = lfsck_namespace_repair_dirent(env, com, dir, obj,
+ info->lti_tmpbuf2, lnr->lnr_name,
+ lnr->lnr_type, true, false);
+ if (rc1 >= 0) {
+ rename_repaired = true;
+ if (rc1 > 0) {
+ ns->ln_dirent_repaired++;
+ rc1 = lfsck_namespace_trace_update(env,
+ com, cfid,
+ LNTF_RECHECK_NAME_HASH, true);
+ }
+ }
+
+ if (rc1 < 0)
+ goto next;
+ }
+
+ if (repair_linkea) {
+ struct lustre_handle lh = { 0 };
+
+ rc1 = linkea_data_new(&ldata, &info->lti_big_buf);
+ if (rc1 != 0)
+ goto next;
+
+ rc1 = linkea_add_buf(&ldata, cname, lfsck_dto2fid(dir));
+ if (rc1 != 0)
+ goto next;
+
+ rc1 = lfsck_ibits_lock(env, lfsck, obj, &lh,
+ MDS_INODELOCK_UPDATE |
+ MDS_INODELOCK_XATTR, LCK_EX);
+ lfsck_ibits_unlock(&lh, LCK_EX);
+ if (rc1 != 0)
+ goto next;
+
+ rc1 = lfsck_namespace_rebuild_linkea(env, com, obj,
+ &ldata);
+ if (rc1 >= 0) {
+ linkea_repaired = true;
+ if (rc1 > 0)
+ ns->ln_linkea_repaired++;
+ }
+ }
+
+next:
+ CDEBUG(D_LFSCK, "%s: namespace LFSCK repair the shard "
+ "%d "DFID" of the striped directory "DFID" with "
+ "dangling %s/%s, rename %s/%s, llinkea %s/%s, "
+ "repair_lmvea %s/%s: rc = %d\n", lfsck_lfsck2name(lfsck),
+ i, PFID(cfid), PFID(&lnr->lnr_fid),
+ create ? "yes" : "no", create_repaired ? "yes" : "no",
+ rename ? "yes" : "no", rename_repaired ? "yes" : "no",
+ repair_linkea ? "yes" : "no",
+ linkea_repaired ? "yes" : "no",
+ repair_lmvea ? "yes" : "no",
+ lmvea_repaired ? "yes" : "no", rc1);
+
+ if (obj != NULL && !IS_ERR(obj))
+ lfsck_object_put(env, obj);
+
+ if (rc1 < 0) {
+ rc = rc1;
+ ns->ln_striped_shards_failed++;
+ }
+ }
+
+ RETURN(rc);
+}
+
+/**
+ * Verify the shard's name entry under the striped directory.
+ *
+ * Before all shards of the striped directory scanned, the LFSCK cannot
+ * know whether the master LMV EA is valid or not, and also cannot know
+ * how to repair an invalid shard exactly. For example, the stripe index
+ * stored in the shard's name does not match the stripe index stored in
+ * the slave LMV EA, then the LFSCK cannot know which one is correct.
+ * If the LFSCK just assumed one is correct, and fixed the other, then
+ * as the LFSCK processing, it may find that the former reparation is
+ * wrong and have to roll back. Unfortunately, if some applications saw
+ * the changes and made further modification based on such changes, then
+ * the roll back is almost impossible.
+ *
+ * To avoid above trouble, the LFSCK will scan the master object of the
+ * striped directory twice, that is NOT the same as normal two-stages
+ * scanning, the double scanning the striped directory will happen both
+ * during the first-stage scanning:
+ *
+ * 1) When the striped directory is opened for scanning, the LFSCK will
+ * iterate each shard in turn, and records its slave LMV EA in the
+ * lfsck_lmv::ll_lslr. In this step, if the 'shard' (may be fake
+ * shard) name does not match the shard naming rule, for example, it
+ * does not contains the shard's FID, or not contains index, then we
+ * can remove the bad name entry directly. But if the name is valid,
+ * but the shard has no slave LMV EA or the slave LMV EA does not
+ * match its name, then we just record related information in the
+ * lfsck_lmv::ll_lslr in RAM.
+ *
+ * 2) When all the known shards have been scanned, then the engine will
+ * generate a dummy request (via lfsck_namespace_close_dir) to tell
+ * the assistant thread that all the known shards have been scanned.
+ * Since the assistant has got the global knowledge about the index
+ * conflict, stripe count, hash type, and so on. Then the assistant
+ * thread will scan the lfsck_lmv::ll_lslr, and for every shard in
+ * the record, check and repair inconsistency.
+ *
+ * Generally, the stripe directory has only several shards, and there
+ * will NOT be a lof of striped directory. So double scanning striped
+ * directory will not much affect the LFSCK performance.
+ *
+ * \param[in] env pointer to the thread context
+ * \param[in] com pointer to the lfsck component
+ * \param[in] dir pointer to the master MDT-object of the
+ * striped directory
+ * \param[in] lnr pointer to the namespace request that contains the
+ * shard's name, parent object, parent's LMV, and ect.
+ *
+ * \retval zero for succeed
+ * \retval negative error number on failure
+ */
+int lfsck_namespace_handle_striped_master(const struct lu_env *env,
+ struct lfsck_component *com,
+ struct dt_object *dir,
+ struct lfsck_namespace_req *lnr)
+{
+ struct lfsck_thread_info *info = lfsck_env_info(env);
+ struct lmv_mds_md_v1 *lmv = &info->lti_lmv;
+ struct lfsck_instance *lfsck = com->lc_lfsck;
+ struct lfsck_namespace *ns = com->lc_file_ram;
+ struct lfsck_lmv *llmv = lnr->lnr_lmv;
+ const struct lu_fid *pfid = lfsck_dto2fid(dir);
+ struct dt_object *obj = NULL;
+ struct dt_device *dev = NULL;
+ int shard_idx = 0;
+ int stripe = 0;
+ int rc = 0;
+ int depth = 0;
+ bool repaired = false;
+ enum lfsck_namespace_inconsistency_type type = LNIT_NONE;
+ ENTRY;
+
+ if (unlikely(llmv->ll_ignore))
+ RETURN(0);
+
+ shard_idx = lfsck_find_mdt_idx_by_fid(env, lfsck, &lnr->lnr_fid);
+ if (shard_idx < 0)
+ GOTO(fail_lmv, rc = shard_idx);
+
+ if (shard_idx == lfsck_dev_idx(lfsck->li_bottom)) {
+ if (unlikely(strcmp(lnr->lnr_name, dotdot) == 0))
+ GOTO(out, rc = 0);
+
+ dev = lfsck->li_next;
+ } else {
+ struct lfsck_tgt_desc *ltd;
+
+ /* Usually, some local filesystem consistency verification
+ * tools can guarantee the local namespace tree consistenct.
+ * So the LFSCK will only verify the remote directory. */
+ if (unlikely(strcmp(lnr->lnr_name, dotdot) == 0)) {
+ rc = lfsck_namespace_trace_update(env, com, pfid,
+ LNTF_CHECK_PARENT, true);
+
+ GOTO(out, rc);
+ }
+
+ ltd = LTD_TGT(&lfsck->li_mdt_descs, shard_idx);
+ if (unlikely(ltd == NULL)) {
+ CDEBUG(D_LFSCK, "%s: cannot talk with MDT %x which "
+ "did not join the namespace LFSCK\n",
+ lfsck_lfsck2name(lfsck), shard_idx);
+ lfsck_lad_set_bitmap(env, com, shard_idx);
+
+ GOTO(fail_lmv, rc = -ENODEV);
+ }
+
+ dev = ltd->ltd_tgt;
+ }
+
+ obj = lfsck_object_find_by_dev_nowait(env, dev, &lnr->lnr_fid);
+ if (IS_ERR(obj)) {
+ if (lfsck_is_dead_obj(dir))
+ RETURN(0);
+
+ GOTO(fail_lmv, rc = PTR_ERR(obj));
+ }
+
+ if (!dt_object_exists(obj)) {
+ stripe = lfsck_shard_name_to_index(env, lnr->lnr_name,
+ lnr->lnr_namelen, lnr->lnr_type, &lnr->lnr_fid);
+ if (stripe < 0) {
+ type = LNIT_BAD_DIRENT;
+
+ GOTO(out, rc = 0);
+ }
+
+dangling:
+ rc = lfsck_namespace_check_exist(env, dir, obj, lnr->lnr_name);
+ if (rc == 0) {
+ memset(lmv, 0, sizeof(*lmv));
+ lmv->lmv_magic = LMV_MAGIC;
+ rc = lfsck_record_lmv(env, com, dir, lnr, lmv, stripe,
+ LSLF_DANGLING, LSLF_NONE, &depth);
+ }
+
+ GOTO(out, rc);
+ }
+
+ stripe = lfsck_shard_name_to_index(env, lnr->lnr_name, lnr->lnr_namelen,
+ lfsck_object_type(obj),
+ &lnr->lnr_fid);
+ if (stripe < 0) {
+ type = LNIT_BAD_DIRENT;
+
+ GOTO(out, rc = 0);
+ }
+
+ rc = lfsck_read_stripe_lmv(env, obj, lmv);
+ if (unlikely(rc == -ENOENT))
+ /* It may happen when the remote object has been removed,
+ * but the local MDT does not aware of that. */
+ goto dangling;
+
+ if (rc == -ENODATA)
+ rc = lfsck_record_lmv(env, com, dir, lnr, lmv, stripe,
+ LSLF_NO_LMVEA, LSLF_NONE, &depth);
+ else if (rc == 0)
+ rc = lfsck_record_lmv(env, com, dir, lnr, lmv, stripe,
+ lmv->lmv_master_mdt_index != stripe ?
+ LSLF_BAD_INDEX1 : LSLF_NONE, LSLF_NONE,
+ &depth);
+
+ GOTO(out, rc);
+
+fail_lmv:
+ llmv->ll_failed = 1;
+
+out:
+ if (rc >= 0 && type == LNIT_NONE && !S_ISDIR(lnr->lnr_type))
+ type = LNIT_BAD_TYPE;
+
+ switch (type) {
+ case LNIT_BAD_TYPE:
+ rc = lfsck_namespace_repair_dirent(env, com, dir, obj,
+ lnr->lnr_name, lnr->lnr_name,
+ lnr->lnr_type, true, false);
+ if (rc > 0)
+ repaired = true;
+ break;
+ case LNIT_BAD_DIRENT:
+ rc = lfsck_namespace_repair_dirent(env, com, dir, obj,
+ lnr->lnr_name, lnr->lnr_name,
+ lnr->lnr_type, false, false);
+ if (rc > 0)
+ repaired = true;
+ break;
+ default:
+ break;
+ }
+
+ if (rc < 0) {
+ CDEBUG(D_LFSCK, "%s: namespace LFSCK assistant fail to handle "
+ "the shard: "DFID", parent "DFID", name %.*s: rc = %d\n",
+ lfsck_lfsck2name(lfsck), PFID(&lnr->lnr_fid),
+ PFID(lfsck_dto2fid(dir)),
+ lnr->lnr_namelen, lnr->lnr_name, rc);
+
+ if ((rc == -ENOTCONN || rc == -ESHUTDOWN || rc == -EREMCHG ||
+ rc == -ETIMEDOUT || rc == -EHOSTDOWN ||
+ rc == -EHOSTUNREACH || rc == -EINPROGRESS) &&
+ dev != NULL && dev != lfsck->li_next)
+ lfsck_lad_set_bitmap(env, com, shard_idx);
+
+ if (!(lfsck->li_bookmark_ram.lb_param & LPF_FAILOUT))
+ rc = 0;
+ } else {
+ if (repaired) {
+ ns->ln_items_repaired++;
+
+ switch (type) {
+ case LNIT_BAD_TYPE:
+ ns->ln_bad_type_repaired++;
+ break;
+ case LNIT_BAD_DIRENT:
+ ns->ln_dirent_repaired++;
+ break;
+ default:
+ break;
+ }
+ }
+
+ rc = 0;
+ }
+
+ if (obj != NULL && !IS_ERR(obj))
+ lfsck_object_put(env, obj);
+
+ return rc;
}