4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 only,
8 * as published by the Free Software Foundation.
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 * GNU General Public License version 2 for more details. A copy is
14 * included in the COPYING file that accompanied this code.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
23 * Copyright (c) 2012, Intel Corporation.
26 * lustre/mdd/mdd_lfsck.c
28 * Top-level entry points into mdd module
30 * LFSCK controller, which scans the whole device through low layer
31 * iteration APIs, drives all lfsck compeonents, controls the speed.
33 * Author: Fan Yong <yong.fan@whamcloud.com>
37 # define EXPORT_SYMTAB
39 #define DEBUG_SUBSYSTEM S_MDS
41 #include <lustre/lustre_idl.h>
42 #include <lustre_fid.h>
43 #include <obd_support.h>
45 #include "mdd_internal.h"
46 #include "mdd_lfsck.h"
48 #define HALF_SEC (CFS_HZ >> 1)
49 #define LFSCK_CHECKPOINT_INTERVAL 60
50 #define MDS_DIR_DUMMY_START 0xffffffffffffffffULL
52 const char lfsck_bookmark_name[] = "lfsck_bookmark";
53 const char lfsck_namespace_name[] = "lfsck_namespace";
57 static inline struct mdd_device *mdd_lfsck2mdd(struct md_lfsck *lfsck)
59 return container_of0(lfsck, struct mdd_device, mdd_lfsck);
62 static inline char *mdd_lfsck2name(struct md_lfsck *lfsck)
64 struct mdd_device *mdd = mdd_lfsck2mdd(lfsck);
66 return mdd2obd_dev(mdd)->obd_name;
69 static inline void mdd_lfsck_component_get(struct lfsck_component *com)
71 atomic_inc(&com->lc_ref);
74 static inline void mdd_lfsck_component_put(const struct lu_env *env,
75 struct lfsck_component *com)
77 if (atomic_dec_and_test(&com->lc_ref)) {
78 if (com->lc_obj != NULL)
79 lu_object_put(env, &com->lc_obj->do_lu);
80 if (com->lc_file_ram != NULL)
81 OBD_FREE(com->lc_file_ram, com->lc_file_size);
82 if (com->lc_file_disk != NULL)
83 OBD_FREE(com->lc_file_disk, com->lc_file_size);
88 static inline struct lfsck_component *
89 __mdd_lfsck_component_find(struct md_lfsck *lfsck, __u16 type, cfs_list_t *list)
91 struct lfsck_component *com;
93 cfs_list_for_each_entry(com, list, lc_link) {
94 if (com->lc_type == type)
100 static void mdd_lfsck_component_cleanup(const struct lu_env *env,
101 struct lfsck_component *com)
103 if (!cfs_list_empty(&com->lc_link))
104 cfs_list_del_init(&com->lc_link);
105 if (!cfs_list_empty(&com->lc_link_dir))
106 cfs_list_del_init(&com->lc_link_dir);
108 mdd_lfsck_component_put(env, com);
111 static void mdd_lfsck_pos_fill(const struct lu_env *env, struct md_lfsck *lfsck,
112 struct lfsck_position *pos, bool oit_processed,
115 const struct dt_it_ops *iops = &lfsck->ml_obj_oit->do_index_ops->dio_it;
117 spin_lock(&lfsck->ml_lock);
118 if (unlikely(lfsck->ml_di_oit == NULL)) {
119 spin_unlock(&lfsck->ml_lock);
120 memset(pos, 0, sizeof(*pos));
124 pos->lp_oit_cookie = iops->store(env, lfsck->ml_di_oit);
126 LASSERT(pos->lp_oit_cookie > 0);
129 pos->lp_oit_cookie--;
131 if (lfsck->ml_di_dir != NULL) {
132 struct dt_object *dto = lfsck->ml_obj_dir;
134 pos->lp_dir_parent = *lu_object_fid(&dto->do_lu);
135 pos->lp_dir_cookie = dto->do_index_ops->dio_it.store(env,
138 LASSERT(pos->lp_dir_cookie != MDS_DIR_DUMMY_START);
140 if (pos->lp_dir_cookie == MDS_DIR_END_OFF)
141 LASSERT(dir_processed);
143 /* For the dir which just to be processed,
144 * lp_dir_cookie will become MDS_DIR_DUMMY_START,
145 * which can be correctly handled by mdd_lfsck_prep. */
147 pos->lp_dir_cookie--;
149 fid_zero(&pos->lp_dir_parent);
150 pos->lp_dir_cookie = 0;
152 spin_unlock(&lfsck->ml_lock);
155 static inline void mdd_lfsck_pos_set_zero(struct lfsck_position *pos)
157 memset(pos, 0, sizeof(*pos));
160 static inline int mdd_lfsck_pos_is_zero(const struct lfsck_position *pos)
162 return pos->lp_oit_cookie == 0 && fid_is_zero(&pos->lp_dir_parent);
165 static inline int mdd_lfsck_pos_is_eq(const struct lfsck_position *pos1,
166 const struct lfsck_position *pos2)
168 if (pos1->lp_oit_cookie < pos2->lp_oit_cookie)
171 if (pos1->lp_oit_cookie > pos2->lp_oit_cookie)
174 if (fid_is_zero(&pos1->lp_dir_parent) &&
175 !fid_is_zero(&pos2->lp_dir_parent))
178 if (!fid_is_zero(&pos1->lp_dir_parent) &&
179 fid_is_zero(&pos2->lp_dir_parent))
182 if (fid_is_zero(&pos1->lp_dir_parent) &&
183 fid_is_zero(&pos2->lp_dir_parent))
186 LASSERT(lu_fid_eq(&pos1->lp_dir_parent, &pos2->lp_dir_parent));
188 if (pos1->lp_dir_cookie < pos2->lp_dir_cookie)
191 if (pos1->lp_dir_cookie > pos2->lp_dir_cookie)
197 static void mdd_lfsck_close_dir(const struct lu_env *env,
198 struct md_lfsck *lfsck)
200 struct dt_object *dir_obj = lfsck->ml_obj_dir;
201 const struct dt_it_ops *dir_iops = &dir_obj->do_index_ops->dio_it;
202 struct dt_it *dir_di = lfsck->ml_di_dir;
204 spin_lock(&lfsck->ml_lock);
205 lfsck->ml_di_dir = NULL;
206 spin_unlock(&lfsck->ml_lock);
208 dir_iops->put(env, dir_di);
209 dir_iops->fini(env, dir_di);
210 lfsck->ml_obj_dir = NULL;
211 lu_object_put(env, &dir_obj->do_lu);
214 static void __mdd_lfsck_set_speed(struct md_lfsck *lfsck, __u32 limit)
216 lfsck->ml_bookmark_ram.lb_speed_limit = limit;
217 if (limit != LFSCK_SPEED_NO_LIMIT) {
218 if (limit > CFS_HZ) {
219 lfsck->ml_sleep_rate = limit / CFS_HZ;
220 lfsck->ml_sleep_jif = 1;
222 lfsck->ml_sleep_rate = 1;
223 lfsck->ml_sleep_jif = CFS_HZ / limit;
226 lfsck->ml_sleep_jif = 0;
227 lfsck->ml_sleep_rate = 0;
231 static void mdd_lfsck_control_speed(struct md_lfsck *lfsck)
233 struct ptlrpc_thread *thread = &lfsck->ml_thread;
234 struct l_wait_info lwi;
236 if (lfsck->ml_sleep_jif > 0 &&
237 lfsck->ml_new_scanned >= lfsck->ml_sleep_rate) {
238 spin_lock(&lfsck->ml_lock);
239 if (likely(lfsck->ml_sleep_jif > 0 &&
240 lfsck->ml_new_scanned >= lfsck->ml_sleep_rate)) {
241 lwi = LWI_TIMEOUT_INTR(lfsck->ml_sleep_jif, NULL,
242 LWI_ON_SIGNAL_NOOP, NULL);
243 spin_unlock(&lfsck->ml_lock);
245 l_wait_event(thread->t_ctl_waitq,
246 !thread_is_running(thread),
248 lfsck->ml_new_scanned = 0;
250 spin_unlock(&lfsck->ml_lock);
255 /* lfsck_bookmark file ops */
257 static void inline mdd_lfsck_bookmark_to_cpu(struct lfsck_bookmark *des,
258 struct lfsck_bookmark *src)
260 des->lb_magic = le32_to_cpu(src->lb_magic);
261 des->lb_version = le16_to_cpu(src->lb_version);
262 des->lb_param = le16_to_cpu(src->lb_param);
263 des->lb_speed_limit = le32_to_cpu(src->lb_speed_limit);
266 static void inline mdd_lfsck_bookmark_to_le(struct lfsck_bookmark *des,
267 struct lfsck_bookmark *src)
269 des->lb_magic = cpu_to_le32(src->lb_magic);
270 des->lb_version = cpu_to_le16(src->lb_version);
271 des->lb_param = cpu_to_le16(src->lb_param);
272 des->lb_speed_limit = cpu_to_le32(src->lb_speed_limit);
275 static int mdd_lfsck_bookmark_load(const struct lu_env *env,
276 struct md_lfsck *lfsck)
279 int len = sizeof(struct lfsck_bookmark);
282 rc = dt_record_read(env, lfsck->ml_bookmark_obj,
283 mdd_buf_get(env, &lfsck->ml_bookmark_disk, len),
286 struct lfsck_bookmark *bm = &lfsck->ml_bookmark_ram;
288 mdd_lfsck_bookmark_to_cpu(bm, &lfsck->ml_bookmark_disk);
289 if (bm->lb_magic != LFSCK_BOOKMARK_MAGIC) {
290 CWARN("%.16s: invalid lfsck_bookmark magic "
291 "0x%x != 0x%x\n", mdd_lfsck2name(lfsck),
292 bm->lb_magic, LFSCK_BOOKMARK_MAGIC);
293 /* Process it as new lfsck_bookmark. */
297 if (rc == -EFAULT && pos == 0)
298 /* return -ENODATA for empty lfsck_bookmark. */
301 CERROR("%.16s: fail to load lfsck_bookmark, "
302 "expected = %d, rc = %d\n",
303 mdd_lfsck2name(lfsck), len, rc);
308 static int mdd_lfsck_bookmark_store(const struct lu_env *env,
309 struct md_lfsck *lfsck)
311 struct mdd_device *mdd = mdd_lfsck2mdd(lfsck);
312 struct thandle *handle;
313 struct dt_object *obj = lfsck->ml_bookmark_obj;
315 int len = sizeof(struct lfsck_bookmark);
319 mdd_lfsck_bookmark_to_le(&lfsck->ml_bookmark_disk,
320 &lfsck->ml_bookmark_ram);
321 handle = dt_trans_create(env, mdd->mdd_bottom);
322 if (IS_ERR(handle)) {
323 rc = PTR_ERR(handle);
324 CERROR("%.16s: fail to create trans for storing "
325 "lfsck_bookmark: %d\n,", mdd_lfsck2name(lfsck), rc);
329 rc = dt_declare_record_write(env, obj, len, 0, handle);
331 CERROR("%.16s: fail to declare trans for storing "
332 "lfsck_bookmark: %d\n,", mdd_lfsck2name(lfsck), rc);
336 rc = dt_trans_start_local(env, mdd->mdd_bottom, handle);
338 CERROR("%.16s: fail to start trans for storing "
339 "lfsck_bookmark: %d\n,", mdd_lfsck2name(lfsck), rc);
343 rc = dt_record_write(env, obj,
344 mdd_buf_get(env, &lfsck->ml_bookmark_disk, len),
347 CERROR("%.16s: fail to store lfsck_bookmark, expected = %d, "
348 "rc = %d\n", mdd_lfsck2name(lfsck), len, rc);
353 dt_trans_stop(env, mdd->mdd_bottom, handle);
357 static int mdd_lfsck_bookmark_init(const struct lu_env *env,
358 struct md_lfsck *lfsck)
360 struct lfsck_bookmark *mb = &lfsck->ml_bookmark_ram;
363 memset(mb, 0, sizeof(mb));
364 mb->lb_magic = LFSCK_BOOKMARK_MAGIC;
365 mb->lb_version = LFSCK_VERSION_V1;
366 mutex_lock(&lfsck->ml_mutex);
367 rc = mdd_lfsck_bookmark_store(env, lfsck);
368 mutex_unlock(&lfsck->ml_mutex);
372 /* lfsck_namespace file ops */
374 static void inline mdd_lfsck_position_to_cpu(struct lfsck_position *des,
375 struct lfsck_position *src)
377 des->lp_oit_cookie = le64_to_cpu(src->lp_oit_cookie);
378 fid_le_to_cpu(&des->lp_dir_parent, &src->lp_dir_parent);
379 des->lp_dir_cookie = le64_to_cpu(src->lp_dir_cookie);
382 static void inline mdd_lfsck_position_to_le(struct lfsck_position *des,
383 struct lfsck_position *src)
385 des->lp_oit_cookie = cpu_to_le64(src->lp_oit_cookie);
386 fid_cpu_to_le(&des->lp_dir_parent, &src->lp_dir_parent);
387 des->lp_dir_cookie = cpu_to_le64(src->lp_dir_cookie);
390 static void inline mdd_lfsck_namespace_to_cpu(struct lfsck_namespace *des,
391 struct lfsck_namespace *src)
393 des->ln_magic = le32_to_cpu(src->ln_magic);
394 des->ln_status = le32_to_cpu(src->ln_status);
395 des->ln_flags = le32_to_cpu(src->ln_flags);
396 des->ln_success_count = le32_to_cpu(src->ln_success_count);
397 des->ln_run_time_phase1 = le32_to_cpu(src->ln_run_time_phase1);
398 des->ln_run_time_phase2 = le32_to_cpu(src->ln_run_time_phase2);
399 des->ln_time_last_complete = le64_to_cpu(src->ln_time_last_complete);
400 des->ln_time_latest_start = le64_to_cpu(src->ln_time_latest_start);
401 des->ln_time_last_checkpoint =
402 le64_to_cpu(src->ln_time_last_checkpoint);
403 mdd_lfsck_position_to_cpu(&des->ln_pos_latest_start,
404 &src->ln_pos_latest_start);
405 mdd_lfsck_position_to_cpu(&des->ln_pos_last_checkpoint,
406 &src->ln_pos_last_checkpoint);
407 mdd_lfsck_position_to_cpu(&des->ln_pos_first_inconsistent,
408 &src->ln_pos_first_inconsistent);
409 des->ln_items_checked = le64_to_cpu(src->ln_items_checked);
410 des->ln_items_repaired = le64_to_cpu(src->ln_items_repaired);
411 des->ln_items_failed = le64_to_cpu(src->ln_items_failed);
412 des->ln_dirs_checked = le64_to_cpu(src->ln_dirs_checked);
413 des->ln_mlinked_checked = le64_to_cpu(src->ln_mlinked_checked);
414 des->ln_objs_checked_phase2 = le64_to_cpu(src->ln_objs_checked_phase2);
415 des->ln_objs_repaired_phase2 =
416 le64_to_cpu(src->ln_objs_repaired_phase2);
417 des->ln_objs_failed_phase2 = le64_to_cpu(src->ln_objs_failed_phase2);
418 des->ln_objs_nlink_repaired = le64_to_cpu(src->ln_objs_nlink_repaired);
419 des->ln_objs_lost_found = le64_to_cpu(src->ln_objs_lost_found);
420 fid_le_to_cpu(&des->ln_fid_latest_scanned_phase2,
421 &src->ln_fid_latest_scanned_phase2);
424 static void inline mdd_lfsck_namespace_to_le(struct lfsck_namespace *des,
425 struct lfsck_namespace *src)
427 des->ln_magic = cpu_to_le32(src->ln_magic);
428 des->ln_status = cpu_to_le32(src->ln_status);
429 des->ln_flags = cpu_to_le32(src->ln_flags);
430 des->ln_success_count = cpu_to_le32(src->ln_success_count);
431 des->ln_run_time_phase1 = cpu_to_le32(src->ln_run_time_phase1);
432 des->ln_run_time_phase2 = cpu_to_le32(src->ln_run_time_phase2);
433 des->ln_time_last_complete = cpu_to_le64(src->ln_time_last_complete);
434 des->ln_time_latest_start = cpu_to_le64(src->ln_time_latest_start);
435 des->ln_time_last_checkpoint =
436 cpu_to_le64(src->ln_time_last_checkpoint);
437 mdd_lfsck_position_to_le(&des->ln_pos_latest_start,
438 &src->ln_pos_latest_start);
439 mdd_lfsck_position_to_le(&des->ln_pos_last_checkpoint,
440 &src->ln_pos_last_checkpoint);
441 mdd_lfsck_position_to_le(&des->ln_pos_first_inconsistent,
442 &src->ln_pos_first_inconsistent);
443 des->ln_items_checked = cpu_to_le64(src->ln_items_checked);
444 des->ln_items_repaired = cpu_to_le64(src->ln_items_repaired);
445 des->ln_items_failed = cpu_to_le64(src->ln_items_failed);
446 des->ln_dirs_checked = cpu_to_le64(src->ln_dirs_checked);
447 des->ln_mlinked_checked = cpu_to_le64(src->ln_mlinked_checked);
448 des->ln_objs_checked_phase2 = cpu_to_le64(src->ln_objs_checked_phase2);
449 des->ln_objs_repaired_phase2 =
450 cpu_to_le64(src->ln_objs_repaired_phase2);
451 des->ln_objs_failed_phase2 = cpu_to_le64(src->ln_objs_failed_phase2);
452 des->ln_objs_nlink_repaired = cpu_to_le64(src->ln_objs_nlink_repaired);
453 des->ln_objs_lost_found = cpu_to_le64(src->ln_objs_lost_found);
454 fid_cpu_to_le(&des->ln_fid_latest_scanned_phase2,
455 &src->ln_fid_latest_scanned_phase2);
459 * \retval +ve: the lfsck_namespace is broken, the caller should reset it.
460 * \retval 0: succeed.
461 * \retval -ve: failed cases.
463 static int mdd_lfsck_namespace_load(const struct lu_env *env,
464 struct lfsck_component *com)
466 int len = com->lc_file_size;
469 rc = dt_xattr_get(env, com->lc_obj,
470 mdd_buf_get(env, com->lc_file_disk, len),
471 XATTR_NAME_LFSCK_NAMESPACE, BYPASS_CAPA);
473 struct lfsck_namespace *ns = com->lc_file_ram;
475 mdd_lfsck_namespace_to_cpu(ns,
476 (struct lfsck_namespace *)com->lc_file_disk);
477 if (ns->ln_magic != LFSCK_NAMESPACE_MAGIC) {
478 CWARN("%.16s: invalid lfsck_namespace magic "
480 mdd_lfsck2name(com->lc_lfsck),
481 ns->ln_magic, LFSCK_NAMESPACE_MAGIC);
486 } else if (rc != -ENODATA) {
487 CERROR("%.16s: fail to load lfsck_namespace, expected = %d, "
488 "rc = %d\n", mdd_lfsck2name(com->lc_lfsck), len, rc);
495 static int mdd_lfsck_namespace_store(const struct lu_env *env,
496 struct lfsck_component *com, bool init)
498 struct dt_object *obj = com->lc_obj;
499 struct md_lfsck *lfsck = com->lc_lfsck;
500 struct mdd_device *mdd = mdd_lfsck2mdd(lfsck);
501 struct thandle *handle;
502 int len = com->lc_file_size;
506 mdd_lfsck_namespace_to_le((struct lfsck_namespace *)com->lc_file_disk,
507 (struct lfsck_namespace *)com->lc_file_ram);
508 handle = dt_trans_create(env, mdd->mdd_bottom);
509 if (IS_ERR(handle)) {
510 rc = PTR_ERR(handle);
511 CERROR("%.16s: fail to create trans for storing "
512 "lfsck_namespace: %d\n,", mdd_lfsck2name(lfsck), rc);
516 rc = dt_declare_xattr_set(env, obj,
517 mdd_buf_get(env, com->lc_file_disk, len),
518 XATTR_NAME_LFSCK_NAMESPACE, 0, handle);
520 CERROR("%.16s: fail to declare trans for storing "
521 "lfsck_namespace: %d\n,", mdd_lfsck2name(lfsck), rc);
525 rc = dt_trans_start_local(env, mdd->mdd_bottom, handle);
527 CERROR("%.16s: fail to start trans for storing "
528 "lfsck_namespace: %d\n,", mdd_lfsck2name(lfsck), rc);
532 rc = dt_xattr_set(env, obj,
533 mdd_buf_get(env, com->lc_file_disk, len),
534 XATTR_NAME_LFSCK_NAMESPACE,
535 init ? LU_XATTR_CREATE : LU_XATTR_REPLACE,
536 handle, BYPASS_CAPA);
538 CERROR("%.16s: fail to store lfsck_namespace, len = %d, "
539 "rc = %d\n", mdd_lfsck2name(lfsck), len, rc);
544 dt_trans_stop(env, mdd->mdd_bottom, handle);
548 static int mdd_lfsck_namespace_init(const struct lu_env *env,
549 struct lfsck_component *com)
551 struct lfsck_namespace *ns = (struct lfsck_namespace *)com->lc_file_ram;
554 memset(ns, 0, sizeof(*ns));
555 ns->ln_magic = LFSCK_NAMESPACE_MAGIC;
556 ns->ln_status = LS_INIT;
557 down_write(&com->lc_sem);
558 rc = mdd_lfsck_namespace_store(env, com, true);
559 up_write(&com->lc_sem);
563 static int mdd_declare_lfsck_namespace_unlink(const struct lu_env *env,
564 struct mdd_device *mdd,
568 struct thandle *handle)
572 rc = dt_declare_delete(env, p, (const struct dt_key *)name, handle);
576 rc = dt_declare_ref_del(env, c, handle);
580 rc = dt_declare_destroy(env, c, handle);
584 static int mdd_lfsck_namespace_unlink(const struct lu_env *env,
585 struct mdd_device *mdd,
586 struct lfsck_component *com)
588 struct mdd_thread_info *info = mdd_env_info(env);
589 struct lu_fid *fid = &info->mti_fid;
590 struct dt_object *child = com->lc_obj;
591 struct dt_object *parent;
592 struct thandle *handle;
596 parent = dt_store_resolve(env, mdd->mdd_bottom, "", fid);
598 RETURN(rc = PTR_ERR(parent));
600 if (dt_try_as_dir(env, parent))
601 GOTO(out, rc = -ENOTDIR);
603 handle = dt_trans_create(env, mdd->mdd_bottom);
605 GOTO(out, rc = PTR_ERR(handle));
607 rc = mdd_declare_lfsck_namespace_unlink(env, mdd, parent, child,
608 lfsck_namespace_name, handle);
612 rc = dt_trans_start_local(env, mdd->mdd_bottom, handle);
616 rc = dt_delete(env, parent, (struct dt_key *)lfsck_namespace_name,
617 handle, BYPASS_CAPA);
621 rc = child->do_ops->do_ref_del(env, child, handle);
623 lu_local_obj_fid(fid, LFSCK_NAMESPACE_OID);
624 rc = dt_insert(env, parent,
625 (const struct dt_rec*)fid,
626 (const struct dt_key *)lfsck_namespace_name,
627 handle, BYPASS_CAPA, 1);
633 rc = dt_destroy(env, child, handle);
635 lu_object_put(env, &child->do_lu);
642 dt_trans_stop(env, mdd->mdd_bottom, handle);
645 lu_object_put(env, &parent->do_lu);
651 static int mdd_lfsck_namespace_reset(const struct lu_env *env,
652 struct lfsck_component *com, bool init)
654 struct mdd_thread_info *info = mdd_env_info(env);
655 struct lu_fid *fid = &info->mti_fid;
656 struct lfsck_namespace *ns = (struct lfsck_namespace *)com->lc_file_ram;
657 struct mdd_device *mdd = mdd_lfsck2mdd(com->lc_lfsck);
658 struct md_object *mdo;
659 struct dt_object *dto;
663 down_write(&com->lc_sem);
665 memset(ns, 0, sizeof(*ns));
667 __u32 count = ns->ln_success_count;
668 __u64 last_time = ns->ln_time_last_complete;
670 memset(ns, 0, sizeof(*ns));
671 ns->ln_success_count = count;
672 ns->ln_time_last_complete = last_time;
674 ns->ln_magic = LFSCK_NAMESPACE_MAGIC;
675 ns->ln_status = LS_INIT;
677 rc = mdd_lfsck_namespace_unlink(env, mdd, com);
681 lu_local_obj_fid(fid, LFSCK_NAMESPACE_OID);
682 mdo = llo_store_create_index(env, &mdd->mdd_md_dev, mdd->mdd_bottom, "",
683 lfsck_namespace_name, fid,
686 GOTO(out, rc = PTR_ERR(mdo));
688 lu_object_put(env, &mdo->mo_lu);
689 dto = dt_store_open(env, mdd->mdd_bottom, "", lfsck_namespace_name, fid);
691 GOTO(out, rc = PTR_ERR(dto));
694 rc = dto->do_ops->do_index_try(env, dto, &dt_lfsck_features);
698 rc = mdd_lfsck_namespace_store(env, com, true);
703 up_write(&com->lc_sem);
708 mdd_lfsck_namespace_fail(const struct lu_env *env, struct lfsck_component *com,
709 bool oit, bool new_checked)
711 struct lfsck_namespace *ns = (struct lfsck_namespace *)com->lc_file_ram;
713 down_write(&com->lc_sem);
715 com->lc_new_checked++;
716 ns->ln_items_failed++;
717 if (mdd_lfsck_pos_is_zero(&ns->ln_pos_first_inconsistent))
718 mdd_lfsck_pos_fill(env, com->lc_lfsck,
719 &ns->ln_pos_first_inconsistent, oit, !oit);
720 up_write(&com->lc_sem);
723 static int mdd_lfsck_namespace_checkpoint(const struct lu_env *env,
724 struct lfsck_component *com,
727 struct md_lfsck *lfsck = com->lc_lfsck;
728 struct lfsck_namespace *ns =
729 (struct lfsck_namespace *)com->lc_file_ram;
732 if (com->lc_new_checked == 0 && !init)
735 down_write(&com->lc_sem);
737 ns->ln_pos_last_checkpoint = lfsck->ml_pos_current;
739 ns->ln_time_last_checkpoint = ns->ln_time_latest_start;
740 ns->ln_pos_latest_start = lfsck->ml_pos_current;
742 ns->ln_run_time_phase1 += cfs_duration_sec(cfs_time_current() +
743 HALF_SEC - lfsck->ml_time_last_checkpoint);
744 ns->ln_time_last_checkpoint = cfs_time_current_sec();
745 ns->ln_items_checked += com->lc_new_checked;
746 com->lc_new_checked = 0;
749 rc = mdd_lfsck_namespace_store(env, com, false);
751 up_write(&com->lc_sem);
755 static int mdd_lfsck_namespace_prep(const struct lu_env *env,
756 struct lfsck_component *com)
758 struct md_lfsck *lfsck = com->lc_lfsck;
759 struct lfsck_namespace *ns =
760 (struct lfsck_namespace *)com->lc_file_ram;
761 struct lfsck_position *pos = &com->lc_pos_start;
763 if (ns->ln_status == LS_COMPLETED) {
766 rc = mdd_lfsck_namespace_reset(env, com, false);
771 down_write(&com->lc_sem);
773 ns->ln_time_latest_start = cfs_time_current_sec();
775 spin_lock(&lfsck->ml_lock);
776 if (ns->ln_flags & LF_SCANNED_ONCE) {
777 if (!lfsck->ml_drop_dryrun ||
778 mdd_lfsck_pos_is_zero(&ns->ln_pos_first_inconsistent)) {
779 ns->ln_status = LS_SCANNING_PHASE2;
780 cfs_list_del_init(&com->lc_link);
781 cfs_list_add_tail(&com->lc_link,
782 &lfsck->ml_list_double_scan);
783 if (!cfs_list_empty(&com->lc_link_dir))
784 cfs_list_del_init(&com->lc_link_dir);
785 mdd_lfsck_pos_set_zero(pos);
787 ns->ln_status = LS_SCANNING_PHASE1;
788 ns->ln_run_time_phase1 = 0;
789 ns->ln_run_time_phase2 = 0;
790 ns->ln_items_checked = 0;
791 ns->ln_items_repaired = 0;
792 ns->ln_items_failed = 0;
793 ns->ln_dirs_checked = 0;
794 ns->ln_mlinked_checked = 0;
795 ns->ln_objs_checked_phase2 = 0;
796 ns->ln_objs_repaired_phase2 = 0;
797 ns->ln_objs_failed_phase2 = 0;
798 ns->ln_objs_nlink_repaired = 0;
799 ns->ln_objs_lost_found = 0;
800 fid_zero(&ns->ln_fid_latest_scanned_phase2);
801 if (cfs_list_empty(&com->lc_link_dir))
802 cfs_list_add_tail(&com->lc_link_dir,
803 &lfsck->ml_list_dir);
804 *pos = ns->ln_pos_first_inconsistent;
807 ns->ln_status = LS_SCANNING_PHASE1;
808 if (cfs_list_empty(&com->lc_link_dir))
809 cfs_list_add_tail(&com->lc_link_dir,
810 &lfsck->ml_list_dir);
811 if (!lfsck->ml_drop_dryrun ||
812 mdd_lfsck_pos_is_zero(&ns->ln_pos_first_inconsistent)) {
813 *pos = ns->ln_pos_last_checkpoint;
814 pos->lp_oit_cookie++;
815 if (!fid_is_zero(&pos->lp_dir_parent)) {
816 if (pos->lp_dir_cookie == MDS_DIR_END_OFF) {
817 fid_zero(&pos->lp_dir_parent);
819 pos->lp_dir_cookie++;
823 *pos = ns->ln_pos_first_inconsistent;
826 spin_unlock(&lfsck->ml_lock);
828 up_write(&com->lc_sem);
832 static int mdd_lfsck_namespace_exec_oit(const struct lu_env *env,
833 struct lfsck_component *com,
834 struct mdd_object *obj)
836 down_write(&com->lc_sem);
837 com->lc_new_checked++;
838 if (S_ISDIR(mdd_object_type(obj)))
839 ((struct lfsck_namespace *)com->lc_file_ram)->ln_dirs_checked++;
840 up_write(&com->lc_sem);
844 /* XXX: to be implemented in other patch. */
845 static int mdd_lfsck_namespace_exec_dir(const struct lu_env *env,
846 struct lfsck_component *com,
847 struct mdd_object *obj,
848 struct lu_dirent *ent)
850 struct lfsck_namespace *ns =
851 (struct lfsck_namespace *)com->lc_file_ram;
852 const struct lu_name *cname;
855 cname = mdd_name_get_const(env, ent->lde_name, ent->lde_namelen);
856 down_write(&com->lc_sem);
857 com->lc_new_checked++;
859 if (ent->lde_attrs & LUDA_UPGRADE) {
860 ns->ln_flags |= LF_UPGRADE;
862 } else if (ent->lde_attrs & LUDA_REPAIR) {
863 ns->ln_flags |= LF_INCONSISTENT;
869 ns->ln_items_repaired += repaired;
870 up_write(&com->lc_sem);
874 static int mdd_lfsck_namespace_post(const struct lu_env *env,
875 struct lfsck_component *com,
878 struct md_lfsck *lfsck = com->lc_lfsck;
879 struct lfsck_namespace *ns =
880 (struct lfsck_namespace *)com->lc_file_ram;
883 down_write(&com->lc_sem);
885 spin_lock(&lfsck->ml_lock);
887 ns->ln_status = LS_SCANNING_PHASE2;
888 ns->ln_flags |= LF_SCANNED_ONCE;
889 ns->ln_flags &= ~LF_UPGRADE;
890 cfs_list_del_init(&com->lc_link);
891 cfs_list_del_init(&com->lc_link_dir);
892 cfs_list_add_tail(&com->lc_link, &lfsck->ml_list_double_scan);
893 } else if (result == 0) {
894 if (lfsck->ml_paused) {
895 ns->ln_status = LS_PAUSED;
897 ns->ln_status = LS_STOPPED;
898 cfs_list_del_init(&com->lc_link);
899 cfs_list_del_init(&com->lc_link_dir);
900 cfs_list_add_tail(&com->lc_link, &lfsck->ml_list_idle);
903 ns->ln_status = LS_FAILED;
904 cfs_list_del_init(&com->lc_link);
905 cfs_list_del_init(&com->lc_link_dir);
906 cfs_list_add_tail(&com->lc_link, &lfsck->ml_list_idle);
908 spin_unlock(&lfsck->ml_lock);
910 ns->ln_run_time_phase1 += cfs_duration_sec(cfs_time_current() +
911 HALF_SEC - lfsck->ml_time_last_checkpoint);
912 ns->ln_time_last_checkpoint = cfs_time_current_sec();
913 ns->ln_items_checked += com->lc_new_checked;
914 com->lc_new_checked = 0;
916 rc = mdd_lfsck_namespace_store(env, com, false);
918 up_write(&com->lc_sem);
922 /* XXX: to be implemented in other patch. */
924 mdd_lfsck_namespace_dump(const struct lu_env *env, struct lfsck_component *com,
930 /* XXX: to be implemented in other patch. */
931 static int mdd_lfsck_namespace_double_scan(const struct lu_env *env,
932 struct lfsck_component *com)
934 struct md_lfsck *lfsck = com->lc_lfsck;
935 struct lfsck_bookmark *bk = &lfsck->ml_bookmark_ram;
936 struct lfsck_namespace *ns =
937 (struct lfsck_namespace *)com->lc_file_ram;
940 down_write(&com->lc_sem);
942 ns->ln_time_last_checkpoint = cfs_time_current_sec();
943 com->lc_new_checked = 0;
946 ns->ln_status = LS_COMPLETED;
947 if (!(bk->lb_param & LPF_DRYRUN))
949 ~(LF_SCANNED_ONCE | LF_INCONSISTENT | LF_UPGRADE);
950 ns->ln_time_last_complete = ns->ln_time_last_checkpoint;
951 ns->ln_success_count++;
953 spin_lock(&lfsck->ml_lock);
954 cfs_list_del_init(&com->lc_link);
955 cfs_list_add_tail(&com->lc_link, &lfsck->ml_list_idle);
956 spin_unlock(&lfsck->ml_lock);
958 rc = mdd_lfsck_namespace_store(env, com, false);
960 up_write(&com->lc_sem);
964 static struct lfsck_operations mdd_lfsck_namespace_ops = {
965 .lfsck_reset = mdd_lfsck_namespace_reset,
966 .lfsck_fail = mdd_lfsck_namespace_fail,
967 .lfsck_checkpoint = mdd_lfsck_namespace_checkpoint,
968 .lfsck_prep = mdd_lfsck_namespace_prep,
969 .lfsck_exec_oit = mdd_lfsck_namespace_exec_oit,
970 .lfsck_exec_dir = mdd_lfsck_namespace_exec_dir,
971 .lfsck_post = mdd_lfsck_namespace_post,
972 .lfsck_dump = mdd_lfsck_namespace_dump,
973 .lfsck_double_scan = mdd_lfsck_namespace_double_scan,
976 /* LFSCK component setup/cleanup functions */
978 static int mdd_lfsck_namespace_setup(const struct lu_env *env,
979 struct md_lfsck *lfsck)
981 struct mdd_device *mdd = mdd_lfsck2mdd(lfsck);
982 struct lfsck_component *com;
983 struct lfsck_namespace *ns;
984 struct dt_object *obj;
992 CFS_INIT_LIST_HEAD(&com->lc_link);
993 CFS_INIT_LIST_HEAD(&com->lc_link_dir);
994 init_rwsem(&com->lc_sem);
995 atomic_set(&com->lc_ref, 1);
996 com->lc_lfsck = lfsck;
997 com->lc_type = LT_NAMESPACE;
998 com->lc_ops = &mdd_lfsck_namespace_ops;
999 com->lc_file_size = sizeof(struct lfsck_namespace);
1000 OBD_ALLOC(com->lc_file_ram, com->lc_file_size);
1001 if (com->lc_file_ram == NULL)
1002 GOTO(out, rc = -ENOMEM);
1004 OBD_ALLOC(com->lc_file_disk, com->lc_file_size);
1005 if (com->lc_file_disk == NULL)
1006 GOTO(out, rc = -ENOMEM);
1008 obj = dt_store_open(env, mdd->mdd_bottom, "", lfsck_namespace_name,
1009 &mdd_env_info(env)->mti_fid);
1011 GOTO(out, rc = PTR_ERR(obj));
1014 rc = obj->do_ops->do_index_try(env, obj, &dt_lfsck_features);
1018 rc = mdd_lfsck_namespace_load(env, com);
1020 rc = mdd_lfsck_namespace_reset(env, com, true);
1021 else if (rc == -ENODATA)
1022 rc = mdd_lfsck_namespace_init(env, com);
1026 ns = (struct lfsck_namespace *)com->lc_file_ram;
1027 switch (ns->ln_status) {
1032 cfs_list_add_tail(&com->lc_link, &lfsck->ml_list_idle);
1035 CERROR("%s: unknown status: %u\n",
1036 mdd_lfsck2name(lfsck), ns->ln_status);
1038 case LS_SCANNING_PHASE1:
1039 case LS_SCANNING_PHASE2:
1040 /* No need to store the status to disk right now.
1041 * If the system crashed before the status stored,
1042 * it will be loaded back when next time. */
1043 ns->ln_status = LS_CRASHED;
1047 cfs_list_add_tail(&com->lc_link, &lfsck->ml_list_scan);
1048 cfs_list_add_tail(&com->lc_link_dir, &lfsck->ml_list_dir);
1056 mdd_lfsck_component_cleanup(env, com);
1060 /* helper functions for framework */
1062 static int object_is_client_visible(const struct lu_env *env,
1063 struct mdd_device *mdd,
1064 struct mdd_object *obj)
1066 struct lu_fid *fid = &mdd_env_info(env)->mti_fid;
1070 LASSERT(S_ISDIR(mdd_object_type(obj)));
1073 if (mdd_is_root(mdd, mdo2fid(obj))) {
1075 mdd_object_put(env, obj);
1079 mdd_read_lock(env, obj, MOR_TGT_CHILD);
1080 if (unlikely(mdd_is_dead_obj(obj))) {
1081 mdd_read_unlock(env, obj);
1083 mdd_object_put(env, obj);
1087 rc = dt_xattr_get(env, mdd_object_child(obj),
1088 mdd_buf_get(env, NULL, 0), XATTR_NAME_LINK,
1090 mdd_read_unlock(env, obj);
1093 mdd_object_put(env, obj);
1097 if (rc < 0 && rc != -ENODATA) {
1099 mdd_object_put(env, obj);
1103 rc = mdd_parent_fid(env, obj, fid);
1105 mdd_object_put(env, obj);
1109 if (unlikely(lu_fid_eq(fid, &mdd->mdd_local_root_fid)))
1112 obj = mdd_object_find(env, mdd, fid);
1115 else if (IS_ERR(obj))
1116 return PTR_ERR(obj);
1118 /* XXX: need more processing for remote object in the future. */
1119 if (!mdd_object_exists(obj) || mdd_object_remote(obj)) {
1120 mdd_object_put(env, obj);
1129 static void mdd_lfsck_unpack_ent(struct lu_dirent *ent)
1131 fid_le_to_cpu(&ent->lde_fid, &ent->lde_fid);
1132 ent->lde_hash = le64_to_cpu(ent->lde_hash);
1133 ent->lde_reclen = le16_to_cpu(ent->lde_reclen);
1134 ent->lde_namelen = le16_to_cpu(ent->lde_namelen);
1135 ent->lde_attrs = le32_to_cpu(ent->lde_attrs);
1137 /* Make sure the name is terminated with '0'.
1138 * The data (type) after ent::lde_name maybe
1139 * broken, but we do not care. */
1140 ent->lde_name[ent->lde_namelen] = 0;
1143 /* LFSCK wrap functions */
1145 static void mdd_lfsck_fail(const struct lu_env *env, struct md_lfsck *lfsck,
1146 bool oit, bool new_checked)
1148 struct lfsck_component *com;
1150 cfs_list_for_each_entry(com, &lfsck->ml_list_scan, lc_link) {
1151 com->lc_ops->lfsck_fail(env, com, oit, new_checked);
1155 static int mdd_lfsck_checkpoint(const struct lu_env *env,
1156 struct md_lfsck *lfsck, bool oit)
1158 struct lfsck_component *com;
1161 if (likely(cfs_time_beforeq(cfs_time_current(),
1162 lfsck->ml_time_next_checkpoint)))
1165 mdd_lfsck_pos_fill(env, lfsck, &lfsck->ml_pos_current, oit, !oit);
1166 cfs_list_for_each_entry(com, &lfsck->ml_list_scan, lc_link) {
1167 rc = com->lc_ops->lfsck_checkpoint(env, com, false);
1172 lfsck->ml_time_last_checkpoint = cfs_time_current();
1173 lfsck->ml_time_next_checkpoint = lfsck->ml_time_last_checkpoint +
1174 cfs_time_seconds(LFSCK_CHECKPOINT_INTERVAL);
1178 static int mdd_lfsck_prep(struct lu_env *env, struct md_lfsck *lfsck)
1180 struct mdd_device *mdd = mdd_lfsck2mdd(lfsck);
1181 struct mdd_object *obj = NULL;
1182 struct dt_object *dt_obj;
1183 struct lfsck_component *com;
1184 struct lfsck_component *next;
1185 struct lfsck_position *pos = NULL;
1186 const struct dt_it_ops *iops =
1187 &lfsck->ml_obj_oit->do_index_ops->dio_it;
1192 LASSERT(lfsck->ml_obj_dir == NULL);
1193 LASSERT(lfsck->ml_di_dir == NULL);
1195 cfs_list_for_each_entry_safe(com, next, &lfsck->ml_list_scan, lc_link) {
1196 com->lc_new_checked = 0;
1197 if (lfsck->ml_bookmark_ram.lb_param & LPF_DRYRUN)
1198 com->lc_journal = 0;
1200 rc = com->lc_ops->lfsck_prep(env, com);
1204 if ((pos == NULL) ||
1205 (!mdd_lfsck_pos_is_zero(&com->lc_pos_start) &&
1206 mdd_lfsck_pos_is_eq(pos, &com->lc_pos_start) > 0))
1207 pos = &com->lc_pos_start;
1210 /* Init otable-based iterator. */
1212 rc = iops->load(env, lfsck->ml_di_oit, 0);
1213 GOTO(out, rc = (rc >= 0 ? 0 : rc));
1216 rc = iops->load(env, lfsck->ml_di_oit, pos->lp_oit_cookie);
1220 if (fid_is_zero(&pos->lp_dir_parent))
1223 /* Find the directory for namespace-based traverse. */
1224 obj = mdd_object_find(env, mdd, &pos->lp_dir_parent);
1227 else if (IS_ERR(obj))
1228 RETURN(PTR_ERR(obj));
1230 /* XXX: need more processing for remote object in the future. */
1231 if (!mdd_object_exists(obj) || mdd_object_remote(obj) ||
1232 unlikely(!S_ISDIR(mdd_object_type(obj))))
1235 if (unlikely(mdd_is_dead_obj(obj)))
1238 dt_obj = mdd_object_child(obj);
1239 if (unlikely(!dt_try_as_dir(env, dt_obj)))
1240 GOTO(out, rc = -ENOTDIR);
1242 /* Init the namespace-based directory traverse. */
1243 iops = &dt_obj->do_index_ops->dio_it;
1244 di = iops->init(env, dt_obj, lfsck->ml_args_dir, BYPASS_CAPA);
1246 GOTO(out, rc = PTR_ERR(di));
1248 rc = iops->load(env, di, pos->lp_dir_cookie);
1250 rc = iops->next(env, di);
1256 iops->fini(env, di);
1260 lfsck->ml_obj_dir = dt_obj;
1261 spin_lock(&lfsck->ml_lock);
1262 lfsck->ml_di_dir = di;
1263 spin_unlock(&lfsck->ml_lock);
1270 mdd_object_put(env, obj);
1273 return (rc > 0 ? 0 : rc);
1275 mdd_lfsck_pos_fill(env, lfsck, &lfsck->ml_pos_current, false, false);
1276 cfs_list_for_each_entry(com, &lfsck->ml_list_scan, lc_link) {
1277 rc = com->lc_ops->lfsck_checkpoint(env, com, true);
1282 lfsck->ml_time_last_checkpoint = cfs_time_current();
1283 lfsck->ml_time_next_checkpoint = lfsck->ml_time_last_checkpoint +
1284 cfs_time_seconds(LFSCK_CHECKPOINT_INTERVAL);
1288 static int mdd_lfsck_exec_oit(const struct lu_env *env, struct md_lfsck *lfsck,
1289 struct mdd_object *obj)
1291 struct lfsck_component *com;
1292 struct dt_object *dt_obj;
1293 const struct dt_it_ops *iops;
1298 LASSERT(lfsck->ml_obj_dir == NULL);
1300 cfs_list_for_each_entry(com, &lfsck->ml_list_scan, lc_link) {
1301 rc = com->lc_ops->lfsck_exec_oit(env, com, obj);
1306 if (!S_ISDIR(mdd_object_type(obj)) ||
1307 cfs_list_empty(&lfsck->ml_list_dir))
1310 rc = object_is_client_visible(env, mdd_lfsck2mdd(lfsck), obj);
1314 if (unlikely(mdd_is_dead_obj(obj)))
1317 dt_obj = mdd_object_child(obj);
1318 if (unlikely(!dt_try_as_dir(env, dt_obj)))
1319 GOTO(out, rc = -ENOTDIR);
1321 iops = &dt_obj->do_index_ops->dio_it;
1322 di = iops->init(env, dt_obj, lfsck->ml_args_dir, BYPASS_CAPA);
1324 GOTO(out, rc = PTR_ERR(di));
1326 rc = iops->load(env, di, 0);
1328 rc = iops->next(env, di);
1334 iops->fini(env, di);
1338 mdd_object_get(obj);
1339 lfsck->ml_obj_dir = dt_obj;
1340 spin_lock(&lfsck->ml_lock);
1341 lfsck->ml_di_dir = di;
1342 spin_unlock(&lfsck->ml_lock);
1348 mdd_lfsck_fail(env, lfsck, false, false);
1349 return (rc > 0 ? 0 : rc);
1352 static int mdd_lfsck_exec_dir(const struct lu_env *env, struct md_lfsck *lfsck,
1353 struct mdd_object *obj, struct lu_dirent *ent)
1355 struct lfsck_component *com;
1358 cfs_list_for_each_entry(com, &lfsck->ml_list_scan, lc_link) {
1359 rc = com->lc_ops->lfsck_exec_dir(env, com, obj, ent);
1366 static int mdd_lfsck_post(const struct lu_env *env, struct md_lfsck *lfsck,
1369 struct lfsck_component *com;
1370 struct lfsck_component *next;
1373 mdd_lfsck_pos_fill(env, lfsck, &lfsck->ml_pos_current, true, true);
1374 cfs_list_for_each_entry_safe(com, next, &lfsck->ml_list_scan, lc_link) {
1375 rc = com->lc_ops->lfsck_post(env, com, result);
1380 lfsck->ml_time_last_checkpoint = cfs_time_current();
1381 lfsck->ml_time_next_checkpoint = lfsck->ml_time_last_checkpoint +
1382 cfs_time_seconds(LFSCK_CHECKPOINT_INTERVAL);
1386 static int mdd_lfsck_double_scan(const struct lu_env *env,
1387 struct md_lfsck *lfsck)
1389 struct lfsck_component *com;
1390 struct lfsck_component *next;
1393 cfs_list_for_each_entry_safe(com, next, &lfsck->ml_list_double_scan,
1395 if (lfsck->ml_bookmark_ram.lb_param & LPF_DRYRUN)
1396 com->lc_journal = 0;
1398 rc = com->lc_ops->lfsck_double_scan(env, com);
1407 static int mdd_lfsck_dir_engine(const struct lu_env *env,
1408 struct md_lfsck *lfsck)
1410 struct mdd_thread_info *info = mdd_env_info(env);
1411 struct mdd_device *mdd = mdd_lfsck2mdd(lfsck);
1412 const struct dt_it_ops *iops =
1413 &lfsck->ml_obj_dir->do_index_ops->dio_it;
1414 struct dt_it *di = lfsck->ml_di_dir;
1415 struct lu_dirent *ent = &info->mti_ent;
1416 struct lu_fid *fid = &info->mti_fid;
1417 struct lfsck_bookmark *bk = &lfsck->ml_bookmark_ram;
1418 struct ptlrpc_thread *thread = &lfsck->ml_thread;
1423 struct mdd_object *child;
1425 lfsck->ml_new_scanned++;
1426 rc = iops->rec(env, di, (struct dt_rec *)ent,
1427 lfsck->ml_args_dir);
1429 mdd_lfsck_fail(env, lfsck, false, true);
1430 if (bk->lb_param & LPF_FAILOUT)
1436 mdd_lfsck_unpack_ent(ent);
1437 if (ent->lde_attrs & LUDA_IGNORE)
1440 *fid = ent->lde_fid;
1441 child = mdd_object_find(env, mdd, fid);
1442 if (child == NULL) {
1444 } else if (IS_ERR(child)) {
1445 mdd_lfsck_fail(env, lfsck, false, true);
1446 if (bk->lb_param & LPF_FAILOUT)
1447 RETURN(PTR_ERR(child));
1452 /* XXX: need more processing for remote object in the future. */
1453 if (mdd_object_exists(child) && !mdd_object_remote(child))
1454 rc = mdd_lfsck_exec_dir(env, lfsck, child, ent);
1455 mdd_object_put(env, child);
1456 if (rc != 0 && bk->lb_param & LPF_FAILOUT)
1460 rc = mdd_lfsck_checkpoint(env, lfsck, false);
1461 if (rc != 0 && bk->lb_param & LPF_FAILOUT)
1465 mdd_lfsck_control_speed(lfsck);
1466 if (unlikely(!thread_is_running(thread)))
1469 rc = iops->next(env, di);
1472 if (rc > 0 && !lfsck->ml_oit_over)
1473 mdd_lfsck_close_dir(env, lfsck);
1478 static int mdd_lfsck_oit_engine(const struct lu_env *env,
1479 struct md_lfsck *lfsck)
1481 struct mdd_thread_info *info = mdd_env_info(env);
1482 struct mdd_device *mdd = mdd_lfsck2mdd(lfsck);
1483 const struct dt_it_ops *iops =
1484 &lfsck->ml_obj_oit->do_index_ops->dio_it;
1485 struct dt_it *di = lfsck->ml_di_oit;
1486 struct lu_fid *fid = &info->mti_fid;
1487 struct lfsck_bookmark *bk = &lfsck->ml_bookmark_ram;
1488 struct ptlrpc_thread *thread = &lfsck->ml_thread;
1493 struct mdd_object *target;
1495 if (lfsck->ml_di_dir != NULL) {
1496 rc = mdd_lfsck_dir_engine(env, lfsck);
1501 if (unlikely(lfsck->ml_oit_over))
1504 lfsck->ml_new_scanned++;
1505 rc = iops->rec(env, di, (struct dt_rec *)fid, 0);
1507 mdd_lfsck_fail(env, lfsck, true, true);
1508 if (bk->lb_param & LPF_FAILOUT)
1514 target = mdd_object_find(env, mdd, fid);
1515 if (target == NULL) {
1517 } else if (IS_ERR(target)) {
1518 mdd_lfsck_fail(env, lfsck, true, true);
1519 if (bk->lb_param & LPF_FAILOUT)
1520 RETURN(PTR_ERR(target));
1525 /* XXX: In fact, low layer otable-based iteration should not
1526 * return agent object. But before LU-2646 resolved, we
1527 * need more processing for agent object. */
1528 if (mdd_object_exists(target) && !mdd_object_remote(target))
1529 rc = mdd_lfsck_exec_oit(env, lfsck, target);
1530 mdd_object_put(env, target);
1531 if (rc != 0 && bk->lb_param & LPF_FAILOUT)
1535 rc = mdd_lfsck_checkpoint(env, lfsck, true);
1536 if (rc != 0 && bk->lb_param & LPF_FAILOUT)
1540 mdd_lfsck_control_speed(lfsck);
1542 rc = iops->next(env, di);
1544 lfsck->ml_oit_over = 1;
1546 if (unlikely(!thread_is_running(thread)))
1548 } while (rc == 0 || lfsck->ml_di_dir != NULL);
1553 static int mdd_lfsck_main(void *args)
1556 struct md_lfsck *lfsck = (struct md_lfsck *)args;
1557 struct ptlrpc_thread *thread = &lfsck->ml_thread;
1558 struct dt_object *oit_obj = lfsck->ml_obj_oit;
1559 const struct dt_it_ops *oit_iops = &oit_obj->do_index_ops->dio_it;
1560 struct dt_it *oit_di;
1564 cfs_daemonize("lfsck");
1565 rc = lu_env_init(&env, LCT_MD_THREAD | LCT_DT_THREAD);
1567 CERROR("%s: LFSCK, fail to init env, rc = %d\n",
1568 mdd_lfsck2name(lfsck), rc);
1572 oit_di = oit_iops->init(&env, oit_obj, lfsck->ml_args_oit, BYPASS_CAPA);
1573 if (IS_ERR(oit_di)) {
1574 rc = PTR_ERR(oit_di);
1575 CERROR("%s: LFSCK, fail to init iteration, rc = %d\n",
1576 mdd_lfsck2name(lfsck), rc);
1580 spin_lock(&lfsck->ml_lock);
1581 lfsck->ml_di_oit = oit_di;
1582 spin_unlock(&lfsck->ml_lock);
1583 rc = mdd_lfsck_prep(&env, lfsck);
1587 CDEBUG(D_LFSCK, "LFSCK entry: oit_flags = 0x%x, dir_flags = 0x%x, "
1588 "oit_cookie = "LPU64", dir_cookie = "LPU64", parent = "DFID
1589 ", pid = %d\n", lfsck->ml_args_oit, lfsck->ml_args_dir,
1590 lfsck->ml_pos_current.lp_oit_cookie,
1591 lfsck->ml_pos_current.lp_dir_cookie,
1592 PFID(&lfsck->ml_pos_current.lp_dir_parent),
1595 spin_lock(&lfsck->ml_lock);
1596 thread_set_flags(thread, SVC_RUNNING);
1597 spin_unlock(&lfsck->ml_lock);
1598 cfs_waitq_broadcast(&thread->t_ctl_waitq);
1600 if (!cfs_list_empty(&lfsck->ml_list_scan) ||
1601 cfs_list_empty(&lfsck->ml_list_double_scan))
1602 rc = mdd_lfsck_oit_engine(&env, lfsck);
1606 CDEBUG(D_LFSCK, "LFSCK exit: oit_flags = 0x%x, dir_flags = 0x%x, "
1607 "oit_cookie = "LPU64", dir_cookie = "LPU64", parent = "DFID
1608 ", pid = %d, rc = %d\n", lfsck->ml_args_oit, lfsck->ml_args_dir,
1609 lfsck->ml_pos_current.lp_oit_cookie,
1610 lfsck->ml_pos_current.lp_dir_cookie,
1611 PFID(&lfsck->ml_pos_current.lp_dir_parent),
1612 cfs_curproc_pid(), rc);
1614 if (lfsck->ml_paused && cfs_list_empty(&lfsck->ml_list_scan))
1615 oit_iops->put(&env, oit_di);
1617 rc = mdd_lfsck_post(&env, lfsck, rc);
1618 if (lfsck->ml_di_dir != NULL)
1619 mdd_lfsck_close_dir(&env, lfsck);
1622 spin_lock(&lfsck->ml_lock);
1623 lfsck->ml_di_oit = NULL;
1624 spin_unlock(&lfsck->ml_lock);
1626 oit_iops->fini(&env, oit_di);
1628 if (!cfs_list_empty(&lfsck->ml_list_double_scan))
1629 rc = mdd_lfsck_double_scan(&env, lfsck);
1634 /* XXX: Purge the pinned objects in the future. */
1640 spin_lock(&lfsck->ml_lock);
1641 thread_set_flags(thread, SVC_STOPPED);
1642 cfs_waitq_broadcast(&thread->t_ctl_waitq);
1643 spin_unlock(&lfsck->ml_lock);
1647 /* external interfaces */
1649 int mdd_lfsck_set_speed(const struct lu_env *env, struct md_lfsck *lfsck,
1654 mutex_lock(&lfsck->ml_mutex);
1655 __mdd_lfsck_set_speed(lfsck, limit);
1656 rc = mdd_lfsck_bookmark_store(env, lfsck);
1657 mutex_unlock(&lfsck->ml_mutex);
1661 int mdd_lfsck_start(const struct lu_env *env, struct md_lfsck *lfsck,
1662 struct lfsck_start *start)
1664 struct lfsck_bookmark *bk = &lfsck->ml_bookmark_ram;
1665 struct ptlrpc_thread *thread = &lfsck->ml_thread;
1666 struct lfsck_component *com;
1667 struct l_wait_info lwi = { 0 };
1674 if (lfsck->ml_obj_oit == NULL)
1677 /* start == NULL means auto trigger paused LFSCK. */
1678 if (start == NULL && cfs_list_empty(&lfsck->ml_list_scan))
1681 mutex_lock(&lfsck->ml_mutex);
1682 spin_lock(&lfsck->ml_lock);
1683 if (!thread_is_init(thread) && !thread_is_stopped(thread)) {
1684 spin_unlock(&lfsck->ml_lock);
1685 mutex_unlock(&lfsck->ml_mutex);
1689 spin_unlock(&lfsck->ml_lock);
1691 lfsck->ml_paused = 0;
1692 lfsck->ml_oit_over = 0;
1693 lfsck->ml_drop_dryrun = 0;
1694 lfsck->ml_new_scanned = 0;
1696 /* For auto trigger. */
1700 start->ls_version = bk->lb_version;
1701 if (start->ls_valid & LSV_SPEED_LIMIT) {
1702 __mdd_lfsck_set_speed(lfsck, start->ls_speed_limit);
1706 if (start->ls_valid & LSV_ERROR_HANDLE) {
1707 valid |= DOIV_ERROR_HANDLE;
1708 if (start->ls_flags & LPF_FAILOUT)
1709 flags |= DOIF_FAILOUT;
1711 if ((start->ls_flags & LPF_FAILOUT) &&
1712 !(bk->lb_param & LPF_FAILOUT)) {
1713 bk->lb_param |= LPF_FAILOUT;
1715 } else if (!(start->ls_flags & LPF_FAILOUT) &&
1716 (bk->lb_param & LPF_FAILOUT)) {
1717 bk->lb_param &= ~LPF_FAILOUT;
1722 if (start->ls_valid & LSV_DRYRUN) {
1723 if ((start->ls_flags & LPF_DRYRUN) &&
1724 !(bk->lb_param & LPF_DRYRUN)) {
1725 bk->lb_param |= LPF_DRYRUN;
1727 } else if (!(start->ls_flags & LPF_DRYRUN) &&
1728 (bk->lb_param & LPF_DRYRUN)) {
1729 bk->lb_param &= ~LPF_DRYRUN;
1730 lfsck->ml_drop_dryrun = 1;
1736 rc = mdd_lfsck_bookmark_store(env, lfsck);
1741 if (start->ls_flags & LPF_RESET)
1742 flags |= DOIF_RESET;
1744 if (start->ls_active != 0) {
1745 struct lfsck_component *next;
1748 if (start->ls_active == LFSCK_TYPES_ALL)
1749 start->ls_active = LFSCK_TYPES_SUPPORTED;
1751 if (start->ls_active & ~LFSCK_TYPES_SUPPORTED) {
1752 start->ls_active &= ~LFSCK_TYPES_SUPPORTED;
1753 GOTO(out, rc = -ENOTSUPP);
1756 cfs_list_for_each_entry_safe(com, next,
1757 &lfsck->ml_list_scan, lc_link) {
1758 if (!(com->lc_type & start->ls_active)) {
1759 rc = com->lc_ops->lfsck_post(env, com, 0);
1765 while (start->ls_active != 0) {
1766 if (type & start->ls_active) {
1767 com = __mdd_lfsck_component_find(lfsck, type,
1768 &lfsck->ml_list_idle);
1770 /* The component status will be updated
1771 * when its prep() is called later by
1772 * the LFSCK main engine. */
1773 cfs_list_del_init(&com->lc_link);
1774 cfs_list_add_tail(&com->lc_link,
1775 &lfsck->ml_list_scan);
1777 start->ls_active &= ~type;
1783 cfs_list_for_each_entry(com, &lfsck->ml_list_scan, lc_link) {
1784 start->ls_active |= com->lc_type;
1785 if (flags & DOIF_RESET) {
1786 rc = com->lc_ops->lfsck_reset(env, com, false);
1793 lfsck->ml_args_dir = LUDA_64BITHASH | LUDA_VERIFY;
1794 if (bk->lb_param & LPF_DRYRUN)
1795 lfsck->ml_args_dir |= LUDA_VERIFY_DRYRUN;
1797 if (bk->lb_param & LPF_FAILOUT) {
1798 valid |= DOIV_ERROR_HANDLE;
1799 flags |= DOIF_FAILOUT;
1802 if (!cfs_list_empty(&lfsck->ml_list_scan))
1803 flags |= DOIF_OUTUSED;
1805 lfsck->ml_args_oit = (flags << DT_OTABLE_IT_FLAGS_SHIFT) | valid;
1806 thread_set_flags(thread, 0);
1807 rc = cfs_create_thread(mdd_lfsck_main, lfsck, 0);
1809 CERROR("%s: cannot start LFSCK thread, rc = %d\n",
1810 mdd_lfsck2name(lfsck), rc);
1812 l_wait_event(thread->t_ctl_waitq,
1813 thread_is_running(thread) ||
1814 thread_is_stopped(thread),
1820 mutex_unlock(&lfsck->ml_mutex);
1821 return (rc < 0 ? rc : 0);
1824 int mdd_lfsck_stop(const struct lu_env *env, struct md_lfsck *lfsck,
1827 struct ptlrpc_thread *thread = &lfsck->ml_thread;
1828 struct l_wait_info lwi = { 0 };
1831 mutex_lock(&lfsck->ml_mutex);
1832 spin_lock(&lfsck->ml_lock);
1833 if (thread_is_init(thread) || thread_is_stopped(thread)) {
1834 spin_unlock(&lfsck->ml_lock);
1835 mutex_unlock(&lfsck->ml_mutex);
1840 lfsck->ml_paused = 1;
1841 thread_set_flags(thread, SVC_STOPPING);
1842 /* The LFSCK thread may be sleeping on low layer wait queue,
1844 if (likely(lfsck->ml_di_oit != NULL))
1845 lfsck->ml_obj_oit->do_index_ops->dio_it.put(env,
1847 spin_unlock(&lfsck->ml_lock);
1849 cfs_waitq_broadcast(&thread->t_ctl_waitq);
1850 l_wait_event(thread->t_ctl_waitq,
1851 thread_is_stopped(thread),
1853 mutex_unlock(&lfsck->ml_mutex);
1858 static const struct lu_fid lfsck_it_fid = { .f_seq = FID_SEQ_LOCAL_FILE,
1859 .f_oid = OTABLE_IT_OID,
1862 int mdd_lfsck_setup(const struct lu_env *env, struct mdd_device *mdd)
1864 struct md_lfsck *lfsck = &mdd->mdd_lfsck;
1865 struct dt_object *obj;
1869 LASSERT(!lfsck->ml_initialized);
1871 lfsck->ml_initialized = 1;
1872 mutex_init(&lfsck->ml_mutex);
1873 spin_lock_init(&lfsck->ml_lock);
1874 CFS_INIT_LIST_HEAD(&lfsck->ml_list_scan);
1875 CFS_INIT_LIST_HEAD(&lfsck->ml_list_dir);
1876 CFS_INIT_LIST_HEAD(&lfsck->ml_list_double_scan);
1877 CFS_INIT_LIST_HEAD(&lfsck->ml_list_idle);
1878 cfs_waitq_init(&lfsck->ml_thread.t_ctl_waitq);
1880 obj = dt_locate(env, mdd->mdd_bottom, &lfsck_it_fid);
1882 RETURN(PTR_ERR(obj));
1884 lfsck->ml_obj_oit = obj;
1885 rc = obj->do_ops->do_index_try(env, obj, &dt_otable_features);
1887 if (rc == -ENOTSUPP)
1893 obj = dt_store_open(env, mdd->mdd_bottom, "", lfsck_bookmark_name,
1894 &mdd_env_info(env)->mti_fid);
1896 RETURN(PTR_ERR(obj));
1898 lfsck->ml_bookmark_obj = obj;
1899 rc = mdd_lfsck_bookmark_load(env, lfsck);
1901 rc = mdd_lfsck_bookmark_init(env, lfsck);
1905 rc = mdd_lfsck_namespace_setup(env, lfsck);
1906 /* XXX: LFSCK components initialization to be added here. */
1911 void mdd_lfsck_cleanup(const struct lu_env *env, struct mdd_device *mdd)
1913 struct md_lfsck *lfsck = &mdd->mdd_lfsck;
1914 struct ptlrpc_thread *thread = &lfsck->ml_thread;
1915 struct lfsck_component *com;
1917 if (!lfsck->ml_initialized)
1920 LASSERT(thread_is_init(thread) || thread_is_stopped(thread));
1922 if (lfsck->ml_obj_oit != NULL) {
1923 lu_object_put(env, &lfsck->ml_obj_oit->do_lu);
1924 lfsck->ml_obj_oit = NULL;
1927 LASSERT(lfsck->ml_obj_dir == NULL);
1929 if (lfsck->ml_bookmark_obj != NULL) {
1930 lu_object_put(env, &lfsck->ml_bookmark_obj->do_lu);
1931 lfsck->ml_bookmark_obj = NULL;
1934 while (!cfs_list_empty(&lfsck->ml_list_scan)) {
1935 com = cfs_list_entry(lfsck->ml_list_scan.next,
1936 struct lfsck_component,
1938 mdd_lfsck_component_cleanup(env, com);
1941 LASSERT(cfs_list_empty(&lfsck->ml_list_dir));
1943 while (!cfs_list_empty(&lfsck->ml_list_double_scan)) {
1944 com = cfs_list_entry(lfsck->ml_list_double_scan.next,
1945 struct lfsck_component,
1947 mdd_lfsck_component_cleanup(env, com);
1950 while (!cfs_list_empty(&lfsck->ml_list_idle)) {
1951 com = cfs_list_entry(lfsck->ml_list_idle.next,
1952 struct lfsck_component,
1954 mdd_lfsck_component_cleanup(env, com);