* in the LICENSE file that accompanied this code).
*
* You should have received a copy of the GNU General Public License
- * version 2 along with this program; If not, see [sun.com URL with a
- * copy of GPLv2].
+ * version 2 along with this program; If not, see
+ * http://www.sun.com/software/products/lustre/docs/GPLv2.pdf
*
* Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
* CA 95054 USA or visit www.sun.com if you need additional information or
#include "mdd_internal.h"
-static struct lu_object_operations mdd_lu_obj_ops;
+static const struct lu_object_operations mdd_lu_obj_ops;
int mdd_la_get(const struct lu_env *env, struct mdd_object *obj,
struct lu_attr *la, struct lustre_capa *capa)
}
if (unlikely(mti->mti_max_cookie == NULL)) {
OBD_ALLOC(mti->mti_max_cookie, max_cookie_size);
- if (unlikely(mti->mti_max_cookie != NULL))
+ if (likely(mti->mti_max_cookie != NULL))
mti->mti_max_cookie_size = max_cookie_size;
}
+ if (likely(mti->mti_max_cookie != NULL))
+ memset(mti->mti_max_cookie, 0, mti->mti_max_cookie_size);
return mti->mti_max_cookie;
}
}
}
-static int mdd_object_init(const struct lu_env *env, struct lu_object *o)
+static int mdd_object_init(const struct lu_env *env, struct lu_object *o,
+ const struct lu_object_conf *_)
{
- struct mdd_device *d = lu2mdd_dev(o->lo_dev);
- struct lu_object *below;
+ struct mdd_device *d = lu2mdd_dev(o->lo_dev);
+ struct lu_object *below;
struct lu_device *under;
ENTRY;
- under = &d->mdd_child->dd_lu_dev;
- below = under->ld_ops->ldo_object_alloc(env, o->lo_header, under);
+ under = &d->mdd_child->dd_lu_dev;
+ below = under->ld_ops->ldo_object_alloc(env, o->lo_header, under);
mdd_pdlock_init(lu2mdd_obj(o));
if (below == NULL)
- RETURN(-ENOMEM);
+ RETURN(-ENOMEM);
lu_object_add(o, below);
RETURN(0);
static void mdd_object_free(const struct lu_env *env, struct lu_object *o)
{
struct mdd_object *mdd = lu2mdd_obj(o);
-
+
lu_object_fini(o);
OBD_FREE_PTR(mdd);
}
-static int mdd_object_print(const struct lu_env *env, void *cookie,
- lu_printer_t p, const struct lu_object *o)
-{
- return (*p)(env, cookie, LUSTRE_MDD_NAME"-object@%p", o);
-}
-
/* orphan handling is here */
-static void mdd_object_delete(const struct lu_env *env,
- struct lu_object *o)
+static void mdd_object_delete(const struct lu_env *env, struct lu_object *o)
{
struct mdd_object *mdd_obj = lu2mdd_obj(o);
struct thandle *handle = NULL;
if (IS_ERR(handle))
CERROR("Cannot get thandle\n");
else {
- mdd_write_lock(env, mdd_obj);
+ mdd_write_lock(env, mdd_obj, MOR_TGT_CHILD);
/* let's remove obj from the orphan list */
__mdd_orphan_del(env, mdd_obj, handle);
mdd_write_unlock(env, mdd_obj);
}
}
-static struct lu_object_operations mdd_lu_obj_ops = {
- .loo_object_init = mdd_object_init,
- .loo_object_start = mdd_object_start,
- .loo_object_free = mdd_object_free,
- .loo_object_print = mdd_object_print,
+static const struct lu_object_operations mdd_lu_obj_ops = {
+ .loo_object_init = mdd_object_init,
+ .loo_object_start = mdd_object_start,
+ .loo_object_free = mdd_object_free,
.loo_object_delete = mdd_object_delete
};
struct mdd_device *d,
const struct lu_fid *f)
{
- struct lu_object *o, *lo;
- struct mdd_object *m;
- ENTRY;
-
- o = lu_object_find(env, mdd2lu_dev(d)->ld_site, f);
- if (IS_ERR(o))
- m = (struct mdd_object *)o;
- else {
- lo = lu_object_locate(o->lo_header, mdd2lu_dev(d)->ld_type);
- /* remote object can't be located and should be put then */
- if (lo == NULL)
- lu_object_put(env, o);
- m = lu2mdd_obj(lo);
- }
- RETURN(m);
+ return md2mdd_obj(md_object_find_slice(env, &d->mdd_md_dev, f));
}
int mdd_get_flags(const struct lu_env *env, struct mdd_object *obj)
int rc;
ENTRY;
- mdd_read_lock(env, mdd_obj);
+ mdd_read_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = __mdd_lmm_get(env, mdd_obj, ma);
mdd_read_unlock(env, mdd_obj);
RETURN(rc);
#ifdef CONFIG_FS_POSIX_ACL
if (rc == 0 && ma->ma_need & MA_ACL_DEF) {
if (S_ISDIR(mdd_object_type(mdd_obj)))
- rc = mdd_acl_def_get(env, mdd_obj, ma);
+ rc = mdd_def_acl_get(env, mdd_obj, ma);
}
#endif
CDEBUG(D_INODE, "after getattr rc = %d, ma_valid = "LPX64"\n",
int needlock = ma->ma_need & (MA_LOV | MA_LMV | MA_ACL_DEF);
if (needlock)
- mdd_read_lock(env, mdd_obj);
+ mdd_read_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = mdd_attr_get_internal(env, mdd_obj, ma);
if (needlock)
mdd_read_unlock(env, mdd_obj);
LASSERT(mdd_object_exists(mdd_obj));
- mdd_read_lock(env, mdd_obj);
+ mdd_read_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = mdo_xattr_get(env, mdd_obj, buf, name,
mdd_object_capa(env, mdd_obj));
mdd_read_unlock(env, mdd_obj);
LASSERT(mdd_object_exists(mdd_obj));
next = mdd_object_child(mdd_obj);
- mdd_read_lock(env, mdd_obj);
+ mdd_read_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = next->do_body_ops->dbo_read(env, next, buf, &pos,
mdd_object_capa(env, mdd_obj));
mdd_read_unlock(env, mdd_obj);
ENTRY;
- mdd_read_lock(env, mdd_obj);
+ mdd_read_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = mdo_xattr_list(env, mdd_obj, buf, mdd_object_capa(env, mdd_obj));
mdd_read_unlock(env, mdd_obj);
needacl = needacl && (attr->la_valid & LA_MODE);
if (needacl)
- mdd_write_lock(env, obj);
+ mdd_write_lock(env, obj, MOR_TGT_CHILD);
rc = mdd_attr_set_internal(env, obj, attr, handle, needacl);
if (needacl)
mdd_write_unlock(env, obj);
needacl = needacl && (attr->la_valid & LA_MODE);
if (needacl)
- mdd_write_lock(env, obj);
+ mdd_write_lock(env, obj, MOR_TGT_CHILD);
rc = mdd_attr_check_set_internal(env, obj, attr, handle, needacl);
if (needacl)
mdd_write_unlock(env, obj);
la->la_valid &= ~LA_ATIME;
RETURN(0);
}
-
+
/* Check if flags change. */
if (la->la_valid & LA_FLAGS) {
unsigned int oldflags = 0;
(LUSTRE_IMMUTABLE_FL | LUSTRE_APPEND_FL);
if ((uc->mu_fsuid != tmp_la->la_uid) &&
- !mdd_capable(uc, CAP_FOWNER))
+ !mdd_capable(uc, CFS_CAP_FOWNER))
RETURN(-EPERM);
/* XXX: the IMMUTABLE and APPEND_ONLY flags can
if (mdd_is_immutable(obj))
oldflags |= LUSTRE_IMMUTABLE_FL;
if (mdd_is_append(obj))
- oldflags |= LUSTRE_APPEND_FL;
+ oldflags |= LUSTRE_APPEND_FL;
if ((oldflags ^ newflags) &&
- !mdd_capable(uc, CAP_LINUX_IMMUTABLE))
+ !mdd_capable(uc, CFS_CAP_LINUX_IMMUTABLE))
RETURN(-EPERM);
if (!S_ISDIR(tmp_la->la_mode))
if ((la->la_valid & (LA_MTIME | LA_ATIME | LA_CTIME)) &&
!(la->la_valid & ~(LA_MTIME | LA_ATIME | LA_CTIME))) {
if ((uc->mu_fsuid != tmp_la->la_uid) &&
- !mdd_capable(uc, CAP_FOWNER)) {
+ !mdd_capable(uc, CFS_CAP_FOWNER)) {
rc = mdd_permission_internal_locked(env, obj, tmp_la,
- MAY_WRITE);
+ MAY_WRITE,
+ MOR_TGT_CHILD);
if (rc)
RETURN(rc);
}
/* Bypass la_vaild == LA_MODE,
* this is for changing file with SUID or SGID. */
if ((la->la_valid & ~LA_MODE) &&
+ !(ma->ma_attr_flags & MDS_PERM_BYPASS) &&
(uc->mu_fsuid != tmp_la->la_uid) &&
- !mdd_capable(uc, CAP_FOWNER))
+ !mdd_capable(uc, CFS_CAP_FOWNER))
RETURN(-EPERM);
if (la->la_mode == (umode_t) -1)
(tmp_la->la_mode & ~S_IALLUGO);
/* Also check the setgid bit! */
- if (!lustre_in_group_p(uc, (la->la_valid & LA_GID) ? la->la_gid :
- tmp_la->la_gid) && !mdd_capable(uc, CAP_FSETID))
+ if (!lustre_in_group_p(uc, (la->la_valid & LA_GID) ?
+ la->la_gid : tmp_la->la_gid) &&
+ !mdd_capable(uc, CFS_CAP_FSETID))
la->la_mode &= ~S_ISGID;
} else {
la->la_mode = tmp_la->la_mode;
la->la_uid = tmp_la->la_uid;
if (((uc->mu_fsuid != tmp_la->la_uid) ||
(la->la_uid != tmp_la->la_uid)) &&
- !mdd_capable(uc, CAP_CHOWN))
+ !mdd_capable(uc, CFS_CAP_CHOWN))
RETURN(-EPERM);
/* If the user or group of a non-directory has been
if (((uc->mu_fsuid != tmp_la->la_uid) ||
((la->la_gid != tmp_la->la_gid) &&
!lustre_in_group_p(uc, la->la_gid))) &&
- !mdd_capable(uc, CAP_CHOWN))
+ !mdd_capable(uc, CFS_CAP_CHOWN))
RETURN(-EPERM);
/* Likewise, if the user or group of a non-directory
(uc->mu_fsuid == tmp_la->la_uid)) &&
!(ma->ma_attr_flags & MDS_PERM_BYPASS)) {
rc = mdd_permission_internal_locked(env, obj,
- tmp_la, MAY_WRITE);
+ tmp_la, MAY_WRITE,
+ MOR_TGT_CHILD);
if (rc)
RETURN(rc);
}
struct llog_cookie *logcookies = NULL;
int rc, lmm_size = 0, cookie_size = 0;
struct lu_attr *la_copy = &mdd_env_info(env)->mti_la_for_fix;
+#ifdef HAVE_QUOTA_SUPPORT
+ struct obd_device *obd = mdd->mdd_obd_dev;
+ struct mds_obd *mds = &obd->u.mds;
+ unsigned int qnids[MAXQUOTAS] = { 0, 0 };
+ unsigned int qoids[MAXQUOTAS] = { 0, 0 };
+ int quota_opc = 0, block_count = 0;
+ int inode_pending = 0, block_pending = 0;
+#endif
ENTRY;
- mdd_txn_param_build(env, mdd, MDD_TXN_ATTR_SET_OP);
+ mdd_setattr_txn_param_build(env, obj, (struct md_attr *)ma,
+ MDD_TXN_ATTR_SET_OP);
handle = mdd_trans_start(env, mdd);
if (IS_ERR(handle))
RETURN(PTR_ERR(handle));
if (rc)
GOTO(cleanup, rc);
+#ifdef HAVE_QUOTA_SUPPORT
+ if (mds->mds_quota && la_copy->la_valid & (LA_UID | LA_GID)) {
+ struct lu_attr *la_tmp = &mdd_env_info(env)->mti_la;
+
+ rc = mdd_la_get(env, mdd_obj, la_tmp, BYPASS_CAPA);
+ if (!rc) {
+ quota_opc = FSFILT_OP_SETATTR;
+ mdd_quota_wrapper(la_copy, qnids);
+ mdd_quota_wrapper(la_tmp, qoids);
+ /* get file quota for new owner */
+ lquota_chkquota(mds_quota_interface_ref, obd,
+ qnids[USRQUOTA], qnids[GRPQUOTA], 1,
+ &inode_pending, NULL, 0);
+ block_count = (la_tmp->la_blocks + 7) >> 3;
+ if (block_count)
+ /* get block quota for new owner */
+ lquota_chkquota(mds_quota_interface_ref, obd,
+ qnids[USRQUOTA],
+ qnids[GRPQUOTA],
+ block_count, &block_pending,
+ NULL, LQUOTA_FLAGS_BLK);
+ }
+ }
+#endif
+
if (la_copy->la_valid & LA_FLAGS) {
rc = mdd_attr_set_internal_locked(env, mdd_obj, la_copy,
handle, 1);
rc = mdd_lov_setattr_async(env, mdd_obj, lmm, lmm_size,
logcookies);
}
+#ifdef HAVE_QUOTA_SUPPORT
+ if (quota_opc) {
+ if (inode_pending)
+ lquota_pending_commit(mds_quota_interface_ref, obd,
+ qnids[USRQUOTA], qnids[GRPQUOTA],
+ 1, 0);
+ if (block_pending)
+ lquota_pending_commit(mds_quota_interface_ref, obd,
+ qnids[USRQUOTA], qnids[GRPQUOTA],
+ block_count, 1);
+ /* Trigger dqrel/dqacq for original owner and new owner.
+ * If failed, the next call for lquota_chkquota will
+ * process it. */
+ lquota_adjust(mds_quota_interface_ref, obd, qnids, qoids, rc,
+ quota_opc);
+ }
+#endif
RETURN(rc);
}
int rc;
ENTRY;
- mdd_write_lock(env, obj);
+ mdd_write_lock(env, obj, MOR_TGT_CHILD);
rc = __mdd_xattr_set(env, obj, buf, name, fl, handle);
mdd_write_unlock(env, obj);
if (rc)
RETURN(rc);
- if ((uc->mu_fsuid != tmp_la->la_uid) && !mdd_capable(uc, CAP_FOWNER))
+ if ((uc->mu_fsuid != tmp_la->la_uid) &&
+ !mdd_capable(uc, CFS_CAP_FOWNER))
RETURN(-EPERM);
RETURN(rc);
if (IS_ERR(handle))
RETURN(PTR_ERR(handle));
- mdd_write_lock(env, mdd_obj);
+ mdd_write_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = mdo_xattr_del(env, mdd_obj, name, handle,
mdd_object_capa(env, mdd_obj));
mdd_write_unlock(env, mdd_obj);
struct mdd_object *mdd_obj = md2mdd_obj(obj);
struct mdd_device *mdd = mdo2mdd(obj);
struct thandle *handle;
+#ifdef HAVE_QUOTA_SUPPORT
+ struct obd_device *obd = mdd->mdd_obd_dev;
+ struct mds_obd *mds = &obd->u.mds;
+ unsigned int qids[MAXQUOTAS] = { 0, 0 };
+ int quota_opc = 0;
+#endif
int rc;
ENTRY;
if (IS_ERR(handle))
RETURN(-ENOMEM);
- mdd_write_lock(env, mdd_obj);
+ mdd_write_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = mdd_unlink_sanity_check(env, NULL, mdd_obj, ma);
if (rc)
GOTO(cleanup, rc);
rc = mdd_finish_unlink(env, mdd_obj, ma, handle);
+#ifdef HAVE_QUOTA_SUPPORT
+ if (mds->mds_quota && ma->ma_valid & MA_INODE &&
+ ma->ma_attr.la_nlink == 0 && mdd_obj->mod_count == 0) {
+ quota_opc = FSFILT_OP_UNLINK_PARTIAL_CHILD;
+ mdd_quota_wrapper(&ma->ma_attr, qids);
+ }
+#endif
+
EXIT;
cleanup:
mdd_write_unlock(env, mdd_obj);
mdd_trans_stop(env, mdd, rc, handle);
+#ifdef HAVE_QUOTA_SUPPORT
+ if (quota_opc)
+ /* Trigger dqrel on the owner of child. If failed,
+ * the next call for lquota_chkquota will process it */
+ lquota_adjust(mds_quota_interface_ref, obd, qids, 0, rc,
+ quota_opc);
+#endif
return rc;
}
struct mdd_object *mdd_obj = md2mdd_obj(obj);
const struct lu_fid *pfid = spec->u.sp_pfid;
struct thandle *handle;
- int rc;
+#ifdef HAVE_QUOTA_SUPPORT
+ struct obd_device *obd = mdd->mdd_obd_dev;
+ struct mds_obd *mds = &obd->u.mds;
+ unsigned int qids[MAXQUOTAS] = { 0, 0 };
+ int quota_opc = 0, block_count = 0;
+ int inode_pending = 0, block_pending = 0;
+#endif
+ int rc = 0;
ENTRY;
+#ifdef HAVE_QUOTA_SUPPORT
+ if (mds->mds_quota) {
+ quota_opc = FSFILT_OP_CREATE_PARTIAL_CHILD;
+ mdd_quota_wrapper(&ma->ma_attr, qids);
+ /* get file quota for child */
+ lquota_chkquota(mds_quota_interface_ref, obd, qids[USRQUOTA],
+ qids[GRPQUOTA], 1, &inode_pending, NULL, 0);
+ switch (ma->ma_attr.la_mode & S_IFMT) {
+ case S_IFLNK:
+ case S_IFDIR:
+ block_count = 2;
+ break;
+ case S_IFREG:
+ block_count = 1;
+ break;
+ }
+ /* get block quota for child */
+ if (block_count)
+ lquota_chkquota(mds_quota_interface_ref, obd,
+ qids[USRQUOTA], qids[GRPQUOTA],
+ block_count, &block_pending, NULL,
+ LQUOTA_FLAGS_BLK);
+ }
+#endif
+
mdd_txn_param_build(env, mdd, MDD_TXN_OBJECT_CREATE_OP);
handle = mdd_trans_start(env, mdd);
if (IS_ERR(handle))
- RETURN(PTR_ERR(handle));
+ GOTO(out_pending, rc = PTR_ERR(handle));
- mdd_write_lock(env, mdd_obj);
+ mdd_write_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = mdd_oc_sanity_check(env, mdd_obj, ma);
if (rc)
GOTO(unlock, rc);
}
EXIT;
unlock:
- mdd_write_unlock(env, mdd_obj);
if (rc == 0)
- rc = mdd_attr_get_internal_locked(env, mdd_obj, ma);
+ rc = mdd_attr_get_internal(env, mdd_obj, ma);
+ mdd_write_unlock(env, mdd_obj);
mdd_trans_stop(env, mdd, rc, handle);
+out_pending:
+#ifdef HAVE_QUOTA_SUPPORT
+ if (quota_opc) {
+ if (inode_pending)
+ lquota_pending_commit(mds_quota_interface_ref, obd,
+ qids[USRQUOTA], qids[GRPQUOTA],
+ 1, 0);
+ if (block_pending)
+ lquota_pending_commit(mds_quota_interface_ref, obd,
+ qids[USRQUOTA], qids[GRPQUOTA],
+ block_count, 1);
+ /* Trigger dqacq on the owner of child. If failed,
+ * the next call for lquota_chkquota will process it. */
+ lquota_adjust(mds_quota_interface_ref, obd, qids, 0, rc,
+ FSFILT_OP_CREATE_PARTIAL_CHILD);
+ }
+#endif
return rc;
}
if (IS_ERR(handle))
RETURN(-ENOMEM);
- mdd_write_lock(env, mdd_obj);
+ mdd_write_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = mdd_link_sanity_check(env, NULL, NULL, mdd_obj);
if (rc == 0)
__mdd_ref_add(env, mdd_obj, handle);
if (uc && ((uc->mu_valid == UCRED_OLD) ||
(uc->mu_valid == UCRED_NEW)) &&
(uc->mu_fsuid != tmp_la->la_uid) &&
- !mdd_capable(uc, CAP_FOWNER))
+ !mdd_capable(uc, CFS_CAP_FOWNER))
RETURN(-EPERM);
}
#endif
struct mdd_object *mdd_obj = md2mdd_obj(obj);
int rc = 0;
- mdd_write_lock(env, mdd_obj);
+ mdd_write_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = mdd_open_sanity_check(env, mdd_obj, flags);
if (rc == 0)
int rc;
struct mdd_object *mdd_obj = md2mdd_obj(obj);
struct thandle *handle;
+#ifdef HAVE_QUOTA_SUPPORT
+ struct obd_device *obd = mdo2mdd(obj)->mdd_obd_dev;
+ struct mds_obd *mds = &obd->u.mds;
+ unsigned int qids[MAXQUOTAS] = { 0, 0 };
+ int quota_opc = 0;
+#endif
ENTRY;
rc = mdd_log_txn_param_build(env, obj, ma, MDD_TXN_UNLINK_OP);
if (IS_ERR(handle))
RETURN(PTR_ERR(handle));
- mdd_write_lock(env, mdd_obj);
+ mdd_write_lock(env, mdd_obj, MOR_TGT_CHILD);
/* release open count */
mdd_obj->mod_count --;
rc = mdd_iattr_get(env, mdd_obj, ma);
- if (rc == 0 && mdd_obj->mod_count == 0 && ma->ma_attr.la_nlink == 0)
+ if (rc == 0 && mdd_obj->mod_count == 0 && ma->ma_attr.la_nlink == 0) {
rc = mdd_object_kill(env, mdd_obj, ma);
- else
+#ifdef HAVE_QUOTA_SUPPORT
+ if (mds->mds_quota) {
+ quota_opc = FSFILT_OP_UNLINK_PARTIAL_CHILD;
+ mdd_quota_wrapper(&ma->ma_attr, qids);
+ }
+#endif
+ } else {
ma->ma_valid &= ~(MA_LOV | MA_COOKIE);
-
+ }
+
mdd_write_unlock(env, mdd_obj);
mdd_trans_stop(env, mdo2mdd(obj), rc, handle);
+#ifdef HAVE_QUOTA_SUPPORT
+ if (quota_opc)
+ /* Trigger dqrel on the owner of child. If failed,
+ * the next call for lquota_chkquota will process it */
+ lquota_adjust(mds_quota_interface_ref, obd, qids, 0, rc,
+ quota_opc);
+#endif
RETURN(rc);
}
}
static int mdd_dir_page_build(const struct lu_env *env, int first,
- void *area, int nob, struct dt_it_ops *iops,
+ void *area, int nob, const struct dt_it_ops *iops,
struct dt_it *it, __u64 *start, __u64 *end,
struct lu_dirent **last)
{
{
struct dt_it *it;
struct dt_object *next = mdd_object_child(obj);
- struct dt_it_ops *iops;
+ const struct dt_it_ops *iops;
struct page *pg;
struct lu_dirent *last = NULL;
int i;
LASSERT(mdd_object_exists(mdd_obj));
- mdd_read_lock(env, mdd_obj);
+ mdd_read_lock(env, mdd_obj, MOR_TGT_CHILD);
rc = mdd_readpage_sanity_check(env, mdd_obj);
if (rc)
GOTO(out_unlock, rc);
return rc;
}
-struct md_object_operations mdd_obj_ops = {
+static int mdd_object_sync(const struct lu_env *env, struct md_object *obj)
+{
+ struct mdd_object *mdd_obj = md2mdd_obj(obj);
+ struct dt_object *next;
+
+ LASSERT(mdd_object_exists(mdd_obj));
+ next = mdd_object_child(mdd_obj);
+ return next->do_ops->do_object_sync(env, next);
+}
+
+const struct md_object_operations mdd_obj_ops = {
.moo_permission = mdd_permission,
.moo_attr_get = mdd_attr_get,
.moo_attr_set = mdd_attr_set,
.moo_close = mdd_close,
.moo_readpage = mdd_readpage,
.moo_readlink = mdd_readlink,
- .moo_capa_get = mdd_capa_get
+ .moo_capa_get = mdd_capa_get,
+ .moo_object_sync = mdd_object_sync,
};