" in log "LPX64"\n", hdr->lrh_index, rec->cur_hdr.lrh_index,
rec->cur_id, rec->cur_endrec, llh->lgh_id.lgl_oid);
- cfs_spin_lock(&mdd->mdd_cl.mc_user_lock);
- mdd->mdd_cl.mc_lastuser = rec->cur_id;
- cfs_spin_unlock(&mdd->mdd_cl.mc_user_lock);
+ spin_lock(&mdd->mdd_cl.mc_user_lock);
+ mdd->mdd_cl.mc_lastuser = rec->cur_id;
+ spin_unlock(&mdd->mdd_cl.mc_user_lock);
- return LLOG_PROC_BREAK;
+ return LLOG_PROC_BREAK;
}
static int llog_changelog_cancel_cb(const struct lu_env *env,
int rc;
mdd->mdd_cl.mc_index = 0;
- cfs_spin_lock_init(&mdd->mdd_cl.mc_lock);
+ spin_lock_init(&mdd->mdd_cl.mc_lock);
mdd->mdd_cl.mc_starttime = cfs_time_current_64();
mdd->mdd_cl.mc_flags = 0; /* off by default */
mdd->mdd_cl.mc_mask = CHANGELOG_DEFMASK;
- cfs_spin_lock_init(&mdd->mdd_cl.mc_user_lock);
+ spin_lock_init(&mdd->mdd_cl.mc_user_lock);
mdd->mdd_cl.mc_lastuser = 0;
rc = mdd_changelog_llog_init(env, mdd);
mdd2obd_dev(mdd)->obd_name);
rc = -ESRCH;
} else {
- cfs_spin_lock(&mdd->mdd_cl.mc_lock);
- mdd->mdd_cl.mc_flags |= CLM_ON;
- cfs_spin_unlock(&mdd->mdd_cl.mc_lock);
+ spin_lock(&mdd->mdd_cl.mc_lock);
+ mdd->mdd_cl.mc_flags |= CLM_ON;
+ spin_unlock(&mdd->mdd_cl.mc_lock);
rc = mdd_changelog_write_header(env, mdd, CLM_START);
- }
- } else if ((on == 0) && ((mdd->mdd_cl.mc_flags & CLM_ON) == CLM_ON)) {
- LCONSOLE_INFO("%s: changelog off\n",mdd2obd_dev(mdd)->obd_name);
+ }
+ } else if ((on == 0) && ((mdd->mdd_cl.mc_flags & CLM_ON) == CLM_ON)) {
+ LCONSOLE_INFO("%s: changelog off\n",mdd2obd_dev(mdd)->obd_name);
rc = mdd_changelog_write_header(env, mdd, CLM_FINI);
- cfs_spin_lock(&mdd->mdd_cl.mc_lock);
- mdd->mdd_cl.mc_flags &= ~CLM_ON;
- cfs_spin_unlock(&mdd->mdd_cl.mc_lock);
- }
- return rc;
+ spin_lock(&mdd->mdd_cl.mc_lock);
+ mdd->mdd_cl.mc_flags &= ~CLM_ON;
+ spin_unlock(&mdd->mdd_cl.mc_lock);
+ }
+ return rc;
}
/** Remove entries with indicies up to and including \a endrec from the
if (ctxt == NULL)
return -ENXIO;
- cfs_spin_lock(&mdd->mdd_cl.mc_lock);
- cur = (long long)mdd->mdd_cl.mc_index;
- cfs_spin_unlock(&mdd->mdd_cl.mc_lock);
+ spin_lock(&mdd->mdd_cl.mc_lock);
+ cur = (long long)mdd->mdd_cl.mc_index;
+ spin_unlock(&mdd->mdd_cl.mc_lock);
if (endrec > cur)
endrec = cur;
rec->cr_hdr.lrh_len = llog_data_len(sizeof(*rec) + rec->cr.cr_namelen);
rec->cr_hdr.lrh_type = CHANGELOG_REC;
rec->cr.cr_time = cl_time();
- cfs_spin_lock(&mdd->mdd_cl.mc_lock);
+ spin_lock(&mdd->mdd_cl.mc_lock);
rec->cr.cr_index = ++mdd->mdd_cl.mc_index;
- cfs_spin_unlock(&mdd->mdd_cl.mc_lock);
+ spin_unlock(&mdd->mdd_cl.mc_lock);
ctxt = llog_get_context(obd, LLOG_CHANGELOG_ORIG_CTXT);
LASSERT(ctxt);
/*
* No permission check is needed.
*/
-static int mdd_maxsize_get(const struct lu_env *env, struct md_device *m,
- int *md_size, int *cookie_size)
-{
- struct mdd_device *mdd = lu2mdd_dev(&m->md_lu_dev);
- ENTRY;
-
- *md_size = mdd_lov_mdsize(env, mdd);
- *cookie_size = mdd_lov_cookiesize(env, mdd);
-
- RETURN(0);
-}
-
static int mdd_init_capa_ctxt(const struct lu_env *env, struct md_device *m,
int mode, unsigned long timeout, __u32 alg,
struct lustre_capa_key *keys)
{
struct mdd_device *mdd = lu2mdd_dev(&m->md_lu_dev);
- struct mds_obd *mds = &mdd2obd_dev(mdd)->u.mds;
int rc;
ENTRY;
/* need barrier for mds_capa_keys access. */
- cfs_down_write(&mds->mds_notify_lock);
- mds->mds_capa_keys = keys;
- cfs_up_write(&mds->mds_notify_lock);
rc = mdd_child_ops(mdd)->dt_init_capa_ctxt(env, mdd->mdd_child, mode,
timeout, alg, keys);
.o_health_check = mdd_obd_health_check
};
-/* context key constructor/destructor: mdd_ucred_key_init, mdd_ucred_key_fini */
-LU_KEY_INIT_FINI(mdd_ucred, struct md_ucred);
-
-static struct lu_context_key mdd_ucred_key = {
- .lct_tags = LCT_SESSION,
- .lct_init = mdd_ucred_key_init,
- .lct_fini = mdd_ucred_key_fini
-};
-
-struct md_ucred *md_ucred(const struct lu_env *env)
-{
- LASSERT(env->le_ses != NULL);
- return lu_context_key_get(env->le_ses, &mdd_ucred_key);
-}
-EXPORT_SYMBOL(md_ucred);
-
/*
* context key constructor/destructor:
* mdd_capainfo_key_init, mdd_capainfo_key_fini
rec->cur_hdr.lrh_len = sizeof(*rec);
rec->cur_hdr.lrh_type = CHANGELOG_USER_REC;
- cfs_spin_lock(&mdd->mdd_cl.mc_user_lock);
- if (mdd->mdd_cl.mc_lastuser == (unsigned int)(-1)) {
- cfs_spin_unlock(&mdd->mdd_cl.mc_user_lock);
- CERROR("Maximum number of changelog users exceeded!\n");
- GOTO(out, rc = -EOVERFLOW);
- }
- *id = rec->cur_id = ++mdd->mdd_cl.mc_lastuser;
- rec->cur_endrec = mdd->mdd_cl.mc_index;
- cfs_spin_unlock(&mdd->mdd_cl.mc_user_lock);
+ spin_lock(&mdd->mdd_cl.mc_user_lock);
+ if (mdd->mdd_cl.mc_lastuser == (unsigned int)(-1)) {
+ spin_unlock(&mdd->mdd_cl.mc_user_lock);
+ CERROR("Maximum number of changelog users exceeded!\n");
+ GOTO(out, rc = -EOVERFLOW);
+ }
+ *id = rec->cur_id = ++mdd->mdd_cl.mc_lastuser;
+ rec->cur_endrec = mdd->mdd_cl.mc_index;
+ spin_unlock(&mdd->mdd_cl.mc_user_lock);
rc = llog_cat_add(env, ctxt->loc_handle, &rec->cur_hdr, NULL, NULL);
data.mcud_minrec = 0;
data.mcud_usercount = 0;
data.mcud_endrec = endrec;
- cfs_spin_lock(&mdd->mdd_cl.mc_lock);
- endrec = mdd->mdd_cl.mc_index;
- cfs_spin_unlock(&mdd->mdd_cl.mc_lock);
+ spin_lock(&mdd->mdd_cl.mc_lock);
+ endrec = mdd->mdd_cl.mc_index;
+ spin_unlock(&mdd->mdd_cl.mc_lock);
if ((data.mcud_endrec == 0) ||
((data.mcud_endrec > endrec) &&
(data.mcud_endrec != MCUD_UNREGISTER)))
}
/* type constructor/destructor: mdd_type_init, mdd_type_fini */
-LU_TYPE_INIT_FINI(mdd, &mdd_thread_key, &mdd_ucred_key, &mdd_capainfo_key);
+LU_TYPE_INIT_FINI(mdd, &mdd_thread_key, &mdd_capainfo_key);
const struct md_device_operations mdd_ops = {
.mdo_statfs = mdd_statfs,
.mdo_root_get = mdd_root_get,
- .mdo_maxsize_get = mdd_maxsize_get,
.mdo_init_capa_ctxt = mdd_init_capa_ctxt,
.mdo_update_capa_key= mdd_update_capa_key,
.mdo_llog_ctxt_get = mdd_llog_ctxt_get,