* Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
* Use is subject to license terms.
*
- * Copyright (c) 2010, 2016, Intel Corporation.
+ * Copyright (c) 2010, 2017, Intel Corporation.
*/
/*
* This file is part of Lustre, http://www.lustre.org/
#define LDLM_DEFAULT_LRU_SIZE (100 * num_online_cpus())
#define LDLM_DEFAULT_MAX_ALIVE 3900 /* 3900 seconds ~65 min */
#define LDLM_CTIME_AGE_LIMIT (10)
+/* if client lock is unused for that time it can be cancelled if any other
+ * client shows interest in that lock, e.g. glimpse is occured. */
+#define LDLM_DIRTY_AGE_LIMIT (10)
#define LDLM_DEFAULT_PARALLEL_AST_LIMIT 1024
/**
* This feature is commonly referred to as lru_resize.
*/
struct ldlm_pool {
- /** Pool proc directory. */
- struct proc_dir_entry *pl_proc_dir;
+ /** Pool debugfs directory. */
+ struct dentry *pl_debugfs_entry;
/** Pool name, must be long enough to hold compound proc entry name. */
char pl_name[100];
/** Lock for protecting SLV/CLV updates. */
struct completion pl_kobj_unregister;
};
-typedef int (*ldlm_res_policy)(struct ldlm_namespace *, struct ldlm_lock **,
- void *req_cookie, enum ldlm_mode mode,
- __u64 flags, void *data);
+typedef int (*ldlm_res_policy)(const struct lu_env *env,
+ struct ldlm_namespace *,
+ struct ldlm_lock **, void *req_cookie,
+ enum ldlm_mode mode, __u64 flags, void *data);
typedef int (*ldlm_cancel_cbt)(struct ldlm_lock *lock);
* of ldlm_[res_]lvbo_[init,update,fill]() functions.
*/
struct ldlm_valblock_ops {
- int (*lvbo_init)(struct ldlm_resource *res);
- int (*lvbo_update)(struct ldlm_resource *res, struct ldlm_lock *lock,
- struct ptlrpc_request *r, int increase);
+ int (*lvbo_init)(const struct lu_env *env, struct ldlm_resource *res);
+ int (*lvbo_update)(const struct lu_env *env, struct ldlm_resource *res,
+ struct ldlm_lock *lock, struct ptlrpc_request *r,
+ int increase);
int (*lvbo_free)(struct ldlm_resource *res);
/* Return size of lvb data appropriate RPC size can be reserved */
int (*lvbo_size)(struct ldlm_lock *lock);
/* Called to fill in lvb data to RPC buffer @buf */
- int (*lvbo_fill)(struct ldlm_lock *lock, void *buf, int buflen);
+ int (*lvbo_fill)(const struct lu_env *env, struct ldlm_lock *lock,
+ void *buf, int *buflen);
};
/**
/** Flag indicating if namespace is on client instead of server */
enum ldlm_side ns_client;
+ /** name of this namespace */
+ char *ns_name;
+
/** Resource hash table for namespace. */
struct cfs_hash *ns_rs_hash;
/** Client side original connect flags supported by server. */
__u64 ns_orig_connect_flags;
- /* namespace proc dir entry */
- struct proc_dir_entry *ns_proc_dir_entry;
+ /* namespace debugfs dir entry */
+ struct dentry *ns_debugfs_entry;
/**
* Position in global namespace list linking all namespaces on
struct list_head ns_unused_list;
/** Number of locks in the LRU list above */
int ns_nr_unused;
+ struct list_head *ns_last_pos;
/**
* Maximum number of locks permitted in the LRU. If 0, means locks
* This allows the client to start caching negative dentries
* for a directory and may save an RPC for a later stat.
*/
- unsigned int ns_ctime_age_limit;
-
+ time64_t ns_ctime_age_limit;
+ /**
+ * Number of seconds since the lock was last used. The client may
+ * cancel the lock limited by this age and flush related data if
+ * any other client shows interest in it doing glimpse request.
+ * This allows to cache stat data locally for such files early.
+ */
+ time64_t ns_dirty_age_limit;
/**
* Used to rate-limit ldlm_namespace_dump calls.
* \see ldlm_namespace_dump. Increased by 10 seconds every time
* it is called.
*/
- cfs_time_t ns_next_dump;
+ time64_t ns_next_dump;
/** "policy" function that does actual lock conflict determination */
ldlm_res_policy ns_policy;
* The resources in this namespace remember contended state during
* \a ns_contention_time, in seconds.
*/
- unsigned ns_contention_time;
+ time64_t ns_contention_time;
/**
* Limit size of contended extent locks, in bytes.
LCF_ASYNC = 0x1, /* Cancel locks asynchronously. */
LCF_LOCAL = 0x2, /* Cancel locks locally, not notifing server */
LCF_BL_AST = 0x4, /* Cancel LDLM_FL_BL_AST locks in the same RPC */
+ LCF_CONVERT = 0x8, /* Try to convert IBITS lock before cancel */
};
struct ldlm_flock {
struct list_head l_lru;
/**
* Linkage to resource's lock queues according to current lock state.
- * (could be granted, waiting or converting)
+ * (could be granted or waiting)
* Protected by lr_lock in struct ldlm_resource.
*/
struct list_head l_res_link;
wait_queue_head_t l_waitq;
/**
- * Seconds. It will be updated if there is any activity related to
- * the lock, e.g. enqueue the lock or send blocking AST.
- */
- time64_t l_last_activity;
-
- /**
* Time, in nanoseconds, last used by e.g. being matched by lock match.
*/
ktime_t l_last_used;
/** Private storage for lock user. Opaque to LDLM. */
void *l_ast_data;
+
+ union {
+ /**
+ * Seconds. It will be updated if there is any activity related to
+ * the lock at client, e.g. enqueue the lock. For server it is the
+ * time when blocking ast was sent.
+ */
+ time64_t l_activity;
+ time64_t l_blast_sent;
+ };
+
/* separate ost_lvb used mostly by Data-on-MDT for now.
* It is introduced to don't mix with layout lock data. */
struct ost_lvb l_ost_lvb;
* under this lock.
* \see ost_rw_prolong_locks
*/
- cfs_time_t l_callback_timeout;
+ time64_t l_callback_timeout;
/** Local PID of process which created this lock. */
__u32 l_pid;
* which is for server. */
#define l_slc_link l_rk_ast
+#define HANDLE_MAP_SIZE ((LMV_MAX_STRIPE_COUNT + 7) >> 3)
+
+struct lustre_handle_array {
+ unsigned int ha_count;
+ /* ha_map is used as bit flag to indicate handle is remote or local */
+ char ha_map[HANDLE_MAP_SIZE];
+ struct lustre_handle ha_handles[0];
+};
+
/**
* LDLM resource description.
* Basically, resource is a representation for a single object.
* @{ */
/** List of locks in granted state */
struct list_head lr_granted;
- /** List of locks waiting to change their granted mode (converted) */
- struct list_head lr_converting;
/**
* List of locks that could not be granted due to conflicts and
* that are waiting for conflicts to go away */
union {
/**
* When the resource was considered as contended,
- * used only on server side. */
- cfs_time_t lr_contention_time;
+ * used only on server side.
+ */
+ time64_t lr_contention_time;
/**
* Associated inode, used only on client side.
*/
static inline char *
ldlm_ns_name(struct ldlm_namespace *ns)
{
- return ns->ns_rs_hash->hs_name;
+ return ns->ns_name;
}
static inline struct ldlm_namespace *
return &lock->l_resource->lr_ns_bucket->nsb_at_estimate;
}
-static inline int ldlm_lvbo_init(struct ldlm_resource *res)
+static inline int ldlm_lvbo_init(const struct lu_env *env,
+ struct ldlm_resource *res)
{
struct ldlm_namespace *ns = ldlm_res_to_ns(res);
int rc = 0;
mutex_unlock(&res->lr_lvb_mutex);
return 0;
}
- rc = ns->ns_lvbo->lvbo_init(res);
+ rc = ns->ns_lvbo->lvbo_init(env, res);
if (rc < 0) {
CDEBUG(D_DLMTRACE, "lvbo_init failed for resource : rc = %d\n",
rc);
return 0;
}
-static inline int ldlm_lvbo_fill(struct ldlm_lock *lock, void *buf, int len)
+static inline int ldlm_lvbo_fill(const struct lu_env *env,
+ struct ldlm_lock *lock, void *buf, int *len)
{
struct ldlm_namespace *ns = ldlm_lock_to_ns(lock);
int rc;
if (ns->ns_lvbo != NULL) {
LASSERT(ns->ns_lvbo->lvbo_fill != NULL);
/* init lvb now if not already */
- rc = ldlm_lvbo_init(lock->l_resource);
+ rc = ldlm_lvbo_init(env, lock->l_resource);
if (rc < 0) {
CERROR("lock %p: delayed lvb init failed (rc %d)",
lock, rc);
return rc;
}
- return ns->ns_lvbo->lvbo_fill(lock, buf, len);
+ return ns->ns_lvbo->lvbo_fill(env, lock, buf, len);
}
return 0;
}
void *ei_cb_gl; /** lock glimpse callback */
void *ei_cbdata; /** Data to be passed into callbacks. */
void *ei_namespace; /** lock namespace **/
+ u64 ei_inodebits; /** lock inode bits **/
unsigned int ei_enq_slave:1; /** whether enqueue slave stripes */
};
* LDLM_PROCESS_RESCAN:
*
* It's used when policy functions are called from ldlm_reprocess_queue() to
- * reprocess the wait & convert list and try to grant locks, blocking ASTs
+ * reprocess the wait list and try to grant locks, blocking ASTs
* have already been sent in this situation, completion ASTs need be sent for
* the locks being granted.
*
* LDLM_PROCESS_ENQUEUE:
*
* It's used when policy functions are called from ldlm_lock_enqueue() to
- * process the wait & convert list for handling an enqueue request, blocking
+ * process the wait list for handling an enqueue request, blocking
* ASTs have not been sent yet, so list of conflicting locks would be
* collected and ASTs sent.
*
* LDLM_PROCESS_RECOVERY:
*
* It's used when policy functions are called from ldlm_reprocess_queue() to
- * reprocess the wait & convert list when recovery done. In case of blocking
+ * reprocess the wait list when recovery done. In case of blocking
* ASTs are lost before recovery, it needs not only to grant locks if
* available, but also send blocking ASTs to the locks doesn't have AST sent
* flag. Completion ASTs need be sent for the locks being granted.
struct ldlm_res_id lpa_resid;
struct ldlm_extent lpa_extent;
enum ldlm_mode lpa_mode;
- int lpa_timeout;
+ time64_t lpa_timeout;
int lpa_locks_cnt;
int lpa_blocks_cnt;
};
* MDT or OST to pass through LDLM requests to LDLM for handling
* @{
*/
-int ldlm_handle_enqueue(struct ptlrpc_request *req, ldlm_completion_callback,
- ldlm_blocking_callback, ldlm_glimpse_callback);
int ldlm_handle_enqueue0(struct ldlm_namespace *ns, struct ptlrpc_request *req,
- const struct ldlm_request *dlm_req,
- const struct ldlm_callback_suite *cbs);
-int ldlm_handle_convert(struct ptlrpc_request *req);
+ const struct ldlm_request *dlm_req,
+ const struct ldlm_callback_suite *cbs);
int ldlm_handle_convert0(struct ptlrpc_request *req,
- const struct ldlm_request *dlm_req);
+ const struct ldlm_request *dlm_req);
int ldlm_handle_cancel(struct ptlrpc_request *req);
int ldlm_request_cancel(struct ptlrpc_request *req,
const struct ldlm_request *dlm_req,
/** @} ldlm_handlers */
void ldlm_revoke_export_locks(struct obd_export *exp);
-unsigned int ldlm_bl_timeout(struct ldlm_lock *lock);
+time64_t ldlm_bl_timeout(struct ldlm_lock *lock);
#endif
int ldlm_del_waiting_lock(struct ldlm_lock *lock);
-int ldlm_refresh_waiting_lock(struct ldlm_lock *lock, int timeout);
+int ldlm_refresh_waiting_lock(struct ldlm_lock *lock, time64_t timeout);
int ldlm_get_ref(void);
void ldlm_put_ref(void);
int ldlm_init_export(struct obd_export *exp);
* Update Lock Value Block Operations (LVBO) on a resource taking into account
* data from request \a r
*/
-static inline int ldlm_lvbo_update(struct ldlm_resource *res,
+static inline int ldlm_lvbo_update(const struct lu_env *env,
+ struct ldlm_resource *res,
struct ldlm_lock *lock,
struct ptlrpc_request *req, int increase)
{
int rc;
/* delayed lvb init may be required */
- rc = ldlm_lvbo_init(res);
+ rc = ldlm_lvbo_init(env, res);
if (rc < 0) {
CERROR("delayed lvb init failed (rc %d)\n", rc);
return rc;
}
if (ns->ns_lvbo && ns->ns_lvbo->lvbo_update)
- return ns->ns_lvbo->lvbo_update(res, lock, req, increase);
+ return ns->ns_lvbo->lvbo_update(env, res, lock, req, increase);
return 0;
}
-static inline int ldlm_res_lvbo_update(struct ldlm_resource *res,
- struct ptlrpc_request *req, int increase)
+static inline int ldlm_res_lvbo_update(const struct lu_env *env,
+ struct ldlm_resource *res,
+ struct ptlrpc_request *req,
+ int increase)
{
- return ldlm_lvbo_update(res, NULL, req, increase);
+ return ldlm_lvbo_update(env, res, NULL, req, increase);
}
int ldlm_error2errno(enum ldlm_error error);
struct lustre_handle *, int unref);
enum ldlm_mode ldlm_revalidate_lock_handle(const struct lustre_handle *lockh,
__u64 *bits);
-struct ldlm_resource *ldlm_lock_convert(struct ldlm_lock *lock,
- enum ldlm_mode new_mode, __u32 *flags);
-void ldlm_lock_downgrade(struct ldlm_lock *lock, enum ldlm_mode new_mode);
+void ldlm_lock_mode_downgrade(struct ldlm_lock *lock, enum ldlm_mode new_mode);
void ldlm_lock_cancel(struct ldlm_lock *lock);
void ldlm_reprocess_all(struct ldlm_resource *res);
void ldlm_reprocess_recovery_done(struct ldlm_namespace *ns);
enum ldlm_side client);
void ldlm_namespace_get(struct ldlm_namespace *ns);
void ldlm_namespace_put(struct ldlm_namespace *ns);
-int ldlm_proc_setup(void);
-#ifdef CONFIG_PROC_FS
-void ldlm_proc_cleanup(void);
-#else
-static inline void ldlm_proc_cleanup(void) {}
-#endif
+
+int ldlm_debugfs_setup(void);
+void ldlm_debugfs_cleanup(void);
+
+static inline void ldlm_svc_get_eopc(const struct ldlm_request *dlm_req,
+ struct lprocfs_stats *srv_stats)
+{
+ int lock_type = 0, op = 0;
+
+ lock_type = dlm_req->lock_desc.l_resource.lr_type;
+
+ switch (lock_type) {
+ case LDLM_PLAIN:
+ op = PTLRPC_LAST_CNTR + LDLM_PLAIN_ENQUEUE;
+ break;
+ case LDLM_EXTENT:
+ op = PTLRPC_LAST_CNTR + LDLM_EXTENT_ENQUEUE;
+ break;
+ case LDLM_FLOCK:
+ op = PTLRPC_LAST_CNTR + LDLM_FLOCK_ENQUEUE;
+ break;
+ case LDLM_IBITS:
+ op = PTLRPC_LAST_CNTR + LDLM_IBITS_ENQUEUE;
+ break;
+ default:
+ op = 0;
+ break;
+ }
+
+ if (op != 0)
+ lprocfs_counter_incr(srv_stats, op);
+
+ return;
+}
/* resource.c - internal */
struct ldlm_resource *ldlm_resource_get(struct ldlm_namespace *ns,
enum ldlm_mode mode, __u64 *flags, void *lvb,
__u32 lvb_len,
const struct lustre_handle *lockh, int rc);
-int ldlm_cli_enqueue_local(struct ldlm_namespace *ns,
+int ldlm_cli_enqueue_local(const struct lu_env *env,
+ struct ldlm_namespace *ns,
const struct ldlm_res_id *res_id,
enum ldlm_type type, union ldlm_policy_data *policy,
enum ldlm_mode mode, __u64 *flags,
void *data, __u32 lvb_len, enum lvb_type lvb_type,
const __u64 *client_cookie,
struct lustre_handle *lockh);
-int ldlm_cli_convert(const struct lustre_handle *lockh, int new_mode,
- __u32 *flags);
+int ldlm_cli_convert(struct ldlm_lock *lock, __u32 *flags);
int ldlm_cli_update_pool(struct ptlrpc_request *req);
int ldlm_cli_cancel(const struct lustre_handle *lockh,
enum ldlm_cancel_flags cancel_flags);
int ldlm_cli_cancel_list(struct list_head *head, int count,
struct ptlrpc_request *req,
enum ldlm_cancel_flags flags);
+
+int ldlm_inodebits_drop(struct ldlm_lock *lock, __u64 to_drop);
+int ldlm_cli_dropbits(struct ldlm_lock *lock, __u64 drop_bits);
+int ldlm_cli_dropbits_list(struct list_head *converts, __u64 drop_bits);
+
/** @} ldlm_cli_api */
/* mds/handler.c */
* There are not used outside of ldlm.
* @{
*/
-int ldlm_pools_recalc(enum ldlm_side client);
int ldlm_pools_init(void);
void ldlm_pools_fini(void);
int ldlm_pool_shrink(struct ldlm_pool *pl, int nr, gfp_t gfp_mask);
void ldlm_pool_fini(struct ldlm_pool *pl);
int ldlm_pool_setup(struct ldlm_pool *pl, int limit);
-int ldlm_pool_recalc(struct ldlm_pool *pl);
+time64_t ldlm_pool_recalc(struct ldlm_pool *pl);
__u32 ldlm_pool_get_lvf(struct ldlm_pool *pl);
__u64 ldlm_pool_get_slv(struct ldlm_pool *pl);
__u64 ldlm_pool_get_clv(struct ldlm_pool *pl);
return ex1->start <= ex2->start && ex1->end >= ex2->end;
}
+int ldlm_inodebits_drop(struct ldlm_lock *lock, __u64 to_drop);
+
#endif
/** @} LDLM */