* GPL HEADER END
*/
/*
- * Copyright 2008 Sun Microsystems, Inc. All rights reserved
+ * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
* Use is subject to license terms.
*/
/*
#define DEBUG_SUBSYSTEM S_MDS
#include <linux/module.h>
+#ifdef HAVE_EXT4_LDISKFS
+#include <ldiskfs/ldiskfs_jbd2.h>
+#else
#include <linux/jbd.h>
+#endif
#include <obd.h>
#include <obd_class.h>
#include <lustre_ver.h>
#include <obd_support.h>
#include <lprocfs_status.h>
+#ifdef HAVE_EXT4_LDISKFS
+#include <ldiskfs/ldiskfs.h>
+#else
#include <linux/ldiskfs_fs.h>
+#endif
#include <lustre_mds.h>
#include <lustre/lustre_idl.h>
* Get default acl EA only.
* Hold read_lock for mdd_obj.
*/
-int mdd_def_acl_get(const struct lu_env *env, struct mdd_object *mdd_obj,
+int mdd_def_acl_get(const struct lu_env *env, struct mdd_object *mdd_obj,
struct md_attr *ma)
{
struct lu_buf *buf;
if (ma->ma_valid & MA_ACL_DEF)
RETURN(0);
-
+
buf = mdd_buf_get(env, ma->ma_acl, ma->ma_acl_size);
rc = mdo_xattr_get(env, mdd_obj, buf, XATTR_NAME_ACL_DEFAULT,
BYPASS_CAPA);
/*
* Hold write_lock for o.
*/
-int mdd_acl_chmod(const struct lu_env *env, struct mdd_object *o, __u32 mode,
+int mdd_acl_chmod(const struct lu_env *env, struct mdd_object *o, __u32 mode,
struct thandle *handle)
{
struct lu_buf *buf;
ENTRY;
- buf = mdd_buf_get(env, mdd_env_info(env)->mti_xattr_buf,
+ buf = mdd_buf_get(env, mdd_env_info(env)->mti_xattr_buf,
sizeof(mdd_env_info(env)->mti_xattr_buf));
-
+
rc = mdo_xattr_get(env, o, buf, XATTR_NAME_ACL_ACCESS, BYPASS_CAPA);
if ((rc == -EOPNOTSUPP) || (rc == -ENODATA))
RETURN(0);
sizeof(posix_acl_xattr_entry);
if (entry_count <= 0)
RETURN(0);
-
+
rc = lustre_posix_acl_chmod_masq(entry, mode, entry_count);
if (rc)
RETURN(rc);
sizeof(posix_acl_xattr_entry);
if (entry_count <= 0)
RETURN(0);
-
- if (S_ISDIR(*mode)) {
- rc = mdo_xattr_set(env, obj, buf, XATTR_NAME_ACL_DEFAULT, 0,
+
+ if (S_ISDIR(*mode)) {
+ rc = mdo_xattr_set(env, obj, buf, XATTR_NAME_ACL_DEFAULT, 0,
handle, BYPASS_CAPA);
if (rc)
RETURN(rc);
- }
+ }
rc = lustre_posix_acl_create_masq(entry, mode, entry_count);
if (rc <= 0)
int rc;
ENTRY;
- buf = mdd_buf_get(env, mdd_env_info(env)->mti_xattr_buf,
+ buf = mdd_buf_get(env, mdd_env_info(env)->mti_xattr_buf,
sizeof(mdd_env_info(env)->mti_xattr_buf));
rc = mdo_xattr_get(env, obj, buf, XATTR_NAME_ACL_ACCESS,
mdd_object_capa(env, obj));
}
int __mdd_permission_internal(const struct lu_env *env, struct mdd_object *obj,
- struct lu_attr *la, int mask, int needlock)
+ struct lu_attr *la, int mask, int role)
{
struct md_ucred *uc = md_ucred(env);
__u32 mode;
mode >>= 6;
} else {
if (mode & S_IRWXG) {
- if (needlock)
- mdd_read_lock(env, obj);
+ if (role != -1)
+ mdd_read_lock(env, obj, role);
rc = mdd_check_acl(env, obj, la, mask);
- if (needlock)
+ if (role != -1)
mdd_read_unlock(env, obj);
if (rc == -EACCES)
goto check_capabilities;
RETURN(-EACCES);
}
-int mdd_permission(const struct lu_env *env,
+int mdd_permission(const struct lu_env *env,
struct md_object *pobj, struct md_object *cobj,
struct md_attr *ma, int mask)
{
MAY_VTX_PART | MAY_VTX_FULL |
MAY_RGETFACL);
- rc = mdd_permission_internal_locked(env, mdd_cobj, NULL, mask);
+ rc = mdd_permission_internal_locked(env, mdd_cobj, NULL, mask,
+ MOR_TGT_CHILD);
if (!rc && (check_create || check_link))
rc = mdd_may_create(env, mdd_cobj, NULL, 1, check_link);
capa->lc_opc);
if (IS_ERR(oc)) {
rc = PTR_ERR(oc);
- } else {
+ } else if (likely(oc != NULL)) {
capa_cpy(capa, oc);
capa_put(oc);
}