-/* -*- mode: c; c-basic-offset: 8; indent-tabs-mode: nil; -*-
- * vim:expandtab:shiftwidth=8:tabstop=8:
- *
+/*
* GPL HEADER START
*
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
* GPL HEADER END
*/
/*
- * Copyright 2008 Sun Microsystems, Inc. All rights reserved
+ * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
* Use is subject to license terms.
*/
/*
+ * Copyright (c) 2012 Whamcloud, Inc.
+ */
+/*
* This file is part of Lustre, http://www.lustre.org/
* Lustre is a trademark of Sun Microsystems, Inc.
*
#ifndef _LUSTRE_MDC_H
#define _LUSTRE_MDC_H
+/** \defgroup mdc mdc
+ *
+ * @{
+ */
+
#ifdef __KERNEL__
# include <linux/fs.h>
# include <linux/dcache.h>
# ifdef CONFIG_FS_POSIX_ACL
-# ifdef HAVE_XATTR_ACL
-# include <linux/xattr_acl.h>
-# endif /*HAVE_XATTR_ACL */
-# ifdef HAVE_LINUX_POSIX_ACL_XATTR_H
-# include <linux/posix_acl_xattr.h>
-# endif /* HAVE_LINUX_POSIX_ACL_XATTR_H */
+# include <linux/posix_acl_xattr.h>
# endif /* CONFIG_FS_POSIX_ACL */
-# ifndef HAVE_VFS_INTENT_PATCHES
# include <linux/lustre_intent.h>
-# endif /* HAVE_VFS_INTENT_PATCHES */
#endif /* __KERNEL__ */
#include <lustre_handles.h>
#include <libcfs/libcfs.h>
struct obd_device;
struct mdc_rpc_lock {
- cfs_semaphore_t rpcl_sem;
- struct lookup_intent *rpcl_it;
+ cfs_mutex_t rpcl_mutex;
+ struct lookup_intent *rpcl_it;
+ int rpcl_fakes;
};
+#define MDC_FAKE_RPCL_IT ((void *)0x2c0012bfUL)
+
static inline void mdc_init_rpc_lock(struct mdc_rpc_lock *lck)
{
- cfs_sema_init(&lck->rpcl_sem, 1);
+ cfs_mutex_init(&lck->rpcl_mutex);
lck->rpcl_it = NULL;
}
static inline void mdc_get_rpc_lock(struct mdc_rpc_lock *lck,
- struct lookup_intent *it)
+ struct lookup_intent *it)
{
- ENTRY;
- if (!it || (it->it_op != IT_GETATTR && it->it_op != IT_LOOKUP)) {
- cfs_down(&lck->rpcl_sem);
- LASSERT(lck->rpcl_it == NULL);
- lck->rpcl_it = it;
- }
+ ENTRY;
+
+ if (it != NULL && (it->it_op == IT_GETATTR || it->it_op == IT_LOOKUP))
+ return;
+
+ /* This would normally block until the existing request finishes.
+ * If fail_loc is set it will block until the regular request is
+ * done, then set rpcl_it to MDC_FAKE_RPCL_IT. Once that is set
+ * it will only be cleared when all fake requests are finished.
+ * Only when all fake requests are finished can normal requests
+ * be sent, to ensure they are recoverable again. */
+ again:
+ cfs_mutex_lock(&lck->rpcl_mutex);
+
+ if (CFS_FAIL_CHECK_QUIET(OBD_FAIL_MDC_RPCS_SEM)) {
+ lck->rpcl_it = MDC_FAKE_RPCL_IT;
+ lck->rpcl_fakes++;
+ cfs_mutex_unlock(&lck->rpcl_mutex);
+ return;
+ }
+
+ /* This will only happen when the CFS_FAIL_CHECK() was
+ * just turned off but there are still requests in progress.
+ * Wait until they finish. It doesn't need to be efficient
+ * in this extremely rare case, just have low overhead in
+ * the common case when it isn't true. */
+ while (unlikely(lck->rpcl_it == MDC_FAKE_RPCL_IT)) {
+ cfs_mutex_unlock(&lck->rpcl_mutex);
+ cfs_schedule_timeout(cfs_time_seconds(1) / 4);
+ goto again;
+ }
+
+ LASSERT(lck->rpcl_it == NULL);
+ lck->rpcl_it = it;
}
static inline void mdc_put_rpc_lock(struct mdc_rpc_lock *lck,
- struct lookup_intent *it)
+ struct lookup_intent *it)
{
- if (!it || (it->it_op != IT_GETATTR && it->it_op != IT_LOOKUP)) {
- LASSERT(it == lck->rpcl_it);
- lck->rpcl_it = NULL;
- cfs_up(&lck->rpcl_sem);
- }
- EXIT;
+ if (it != NULL && (it->it_op == IT_GETATTR || it->it_op == IT_LOOKUP))
+ goto out;
+
+ if (lck->rpcl_it == MDC_FAKE_RPCL_IT) { /* OBD_FAIL_MDC_RPCS_SEM */
+ cfs_mutex_lock(&lck->rpcl_mutex);
+
+ LASSERTF(lck->rpcl_fakes > 0, "%d\n", lck->rpcl_fakes);
+ lck->rpcl_fakes--;
+
+ if (lck->rpcl_fakes == 0)
+ lck->rpcl_it = NULL;
+
+ } else {
+ LASSERTF(it == lck->rpcl_it, "%p != %p\n", it, lck->rpcl_it);
+ lck->rpcl_it = NULL;
+ }
+
+ cfs_mutex_unlock(&lck->rpcl_mutex);
+ out:
+ EXIT;
}
static inline void mdc_update_max_ea_from_body(struct obd_export *exp,
int mdc_sendpage(struct obd_export *exp, const struct lu_fid *fid,
const struct page *page, int offset);
#endif
+
+/** @} mdc */
+
#endif