4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 only,
8 * as published by the Free Software Foundation.
10 * This program is distributed in the hope that it will be useful, but
11 * WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * General Public License version 2 for more details (a copy is included
14 * in the LICENSE file that accompanied this code).
16 * You should have received a copy of the GNU General Public License
17 * version 2 along with this program; If not, see
18 * http://www.gnu.org/licenses/gpl-2.0.html
23 * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
24 * Use is subject to license terms.
26 * Copyright (c) 2011, 2017, Intel Corporation.
29 * This file is part of Lustre, http://www.lustre.org/
31 * lustre/mdd/mdd_trans.c
33 * Lustre Metadata Server (mdd) routines
35 * Author: Wang Di <wangdi@clusterfs.com>
38 #define DEBUG_SUBSYSTEM S_MDS
40 #include <linux/kthread.h>
42 #include <obd_class.h>
43 #include <lprocfs_status.h>
44 #include <lustre_mds.h>
45 #include <lustre_barrier.h>
47 #include "mdd_internal.h"
49 struct thandle *mdd_trans_create(const struct lu_env *env,
50 struct mdd_device *mdd)
53 struct lu_ucred *uc = lu_ucred_check(env);
55 /* If blocked by the write barrier, then return "-EINPROGRESS"
56 * to the caller. Usually, such error will be forwarded to the
57 * client, and the expected behaviour is to re-try such modify
58 * RPC some time later until the barrier is thawed or expired. */
59 if (unlikely(!barrier_entry(mdd->mdd_bottom)))
60 return ERR_PTR(-EINPROGRESS);
62 th = mdd_child_ops(mdd)->dt_trans_create(env, mdd->mdd_child);
63 if (!IS_ERR(th) && uc)
64 th->th_ignore_quota = !!cap_raised(uc->uc_cap, CAP_SYS_RESOURCE);
69 int mdd_trans_start(const struct lu_env *env, struct mdd_device *mdd,
72 return mdd_child_ops(mdd)->dt_trans_start(env, mdd->mdd_child, th);
75 struct mdd_changelog_gc {
76 struct mdd_device *mcgc_mdd;
80 char mcgc_name[CHANGELOG_USER_NAMELEN_FULL];
83 /* return first registered ChangeLog user idle since too long
84 * use ChangeLog's user plain LLOG mtime for this */
85 static int mdd_changelog_gc_cb(const struct lu_env *env,
86 struct llog_handle *llh,
87 struct llog_rec_hdr *hdr, void *data)
89 struct llog_changelog_user_rec2 *rec;
90 struct mdd_changelog_gc *mcgc = data;
91 struct mdd_device *mdd = mcgc->mcgc_mdd;
95 if ((llh->lgh_hdr->llh_flags & LLOG_F_IS_PLAIN) == 0)
98 rec = container_of(hdr, typeof(*rec), cur_hdr);
100 if (rec->cur_endrec < mcgc->mcgc_minrec &&
101 (mdd->mdd_changelog_emrg_gc ||
102 mdd_changelog_is_too_idle(mdd, rec->cur_endrec, rec->cur_time))) {
103 mcgc->mcgc_mintime = rec->cur_time;
104 mcgc->mcgc_minrec = rec->cur_endrec;
105 mcgc->mcgc_id = rec->cur_id;
106 mdd_chlg_username(rec, mcgc->mcgc_name,
107 sizeof(mcgc->mcgc_name));
112 /* recover space from long-term inactive ChangeLog users */
113 static int mdd_chlg_garbage_collect(void *data)
115 struct mdd_device *mdd = data;
116 struct lu_env *env = NULL;
118 struct llog_ctxt *ctxt;
122 mdd->mdd_cl.mc_gc_task = current;
124 CDEBUG(D_HA, "%s: ChangeLog garbage collect thread start with PID %d\n",
125 mdd2obd_dev(mdd)->obd_name, current->pid);
129 GOTO(out, rc = -ENOMEM);
131 rc = lu_env_init(env, LCT_MD_THREAD);
135 ctxt = llog_get_context(mdd2obd_dev(mdd),
136 LLOG_CHANGELOG_USER_ORIG_CTXT);
138 GOTO(out_env, rc = -ENXIO);
139 if (!(ctxt->loc_handle->lgh_hdr->llh_flags & LLOG_F_IS_CAT))
140 GOTO(out_ctxt, rc = -ENXIO);
143 __u32 time_now = (__u32)ktime_get_real_seconds();
144 struct mdd_changelog_gc mcgc = {
146 .mcgc_minrec = mdd->mdd_cl.mc_index,
150 rc = llog_cat_process(env, ctxt->loc_handle,
151 mdd_changelog_gc_cb, &mcgc, 0, 0);
155 if (!mcgc.mcgc_name[0])
158 CWARN("%s: force deregister of changelog user %s idle for %us with %llu unprocessed records\n",
159 mdd2obd_dev(mdd)->obd_name, mcgc.mcgc_name,
160 time_now - mcgc.mcgc_mintime,
161 mdd->mdd_cl.mc_index - mcgc.mcgc_minrec);
163 mdd_changelog_user_purge(env, mdd, mcgc.mcgc_id);
165 if (mdd->mdd_changelog_emrg_gc &&
166 mdd_changelog_is_space_safe(env, mdd, ctxt->loc_handle, 0))
167 mdd->mdd_changelog_emrg_gc = false;
169 if (kthread_should_stop())
170 GOTO(out_ctxt, rc = 0);
180 spin_lock(&mdd->mdd_cl.mc_lock);
181 mdd->mdd_cl.mc_gc_task = MDD_CHLG_GC_NONE;
182 spin_unlock(&mdd->mdd_cl.mc_lock);
187 int mdd_trans_stop(const struct lu_env *env, struct mdd_device *mdd,
188 int result, struct thandle *handle)
192 handle->th_result = result;
193 rc = mdd_child_ops(mdd)->dt_trans_stop(env, mdd->mdd_child, handle);
194 barrier_exit(mdd->mdd_bottom);
196 /* bottom half of changelog garbage-collection mechanism, started
197 * from mdd_changelog_store(). This is required, as running a
198 * kthead can't occur during a journal transaction is being filled
199 * because otherwise a deadlock can happen if memory reclaim is
200 * triggered by kthreadd when forking the new thread, and thus
201 * I/Os could be attempted to the same device from shrinkers
202 * requiring a new journal transaction to be started when current
203 * could never complete (LU-10680).
205 if (unlikely(mdd->mdd_cl.mc_flags & CLM_ON &&
206 cmpxchg(&mdd->mdd_cl.mc_gc_task, MDD_CHLG_GC_NEED,
207 MDD_CHLG_GC_START) == MDD_CHLG_GC_NEED)) {
208 /* XXX we may want to cmpxchg() only if MDD_CHLG_GC_NEED
209 * to save its cost in the frequent case and have an extra
210 * if/test cost in the rare case where we need to spawn?
212 struct task_struct *gc_task;
213 struct obd_device *obd = mdd2obd_dev(mdd);
215 gc_task = kthread_run(mdd_chlg_garbage_collect, mdd,
217 if (IS_ERR(gc_task)) {
218 CERROR("%s: cannot start ChangeLog garbage collection "
219 "thread: rc = %ld\n", obd->obd_name,
221 mdd->mdd_cl.mc_gc_task = MDD_CHLG_GC_NONE;
223 CDEBUG(D_HA, "%s: a ChangeLog garbage collection "
224 "thread has been started\n", obd->obd_name);
228 /* if operation failed, return \a result, otherwise return status of