* Copyright (c) 2003, 2010, Oracle and/or its affiliates. All rights reserved.
* Use is subject to license terms.
*
- * Copyright (c) 2011, 2012, Whamcloud, Inc.
+ * Copyright (c) 2012, 2013, Intel Corporation.
*/
/*
* This file is part of Lustre, http://www.lustre.org/
#define _OFD_INTERNAL_H
#include <obd.h>
+#include <obd_class.h>
#include <dt_object.h>
+#include <md_object.h>
+#include <lustre_fid.h>
+#include <obd_ost.h>
+#include <lustre_capa.h>
+
+#define OFD_INIT_OBJID 0
+#define OFD_ROCOMPAT_SUPP (0)
+#define OFD_INCOMPAT_SUPP (OBD_INCOMPAT_GROUPS | OBD_INCOMPAT_OST | \
+ OBD_INCOMPAT_COMMON_LR)
+#define OFD_PRECREATE_BATCH_DEFAULT (FILTER_SUBDIR_COUNT * 4)
+
+/* on small filesystems we should not precreate too many objects in
+ * a single transaction, otherwise we can overflow transactions */
+#define OFD_PRECREATE_SMALL_FS (1024ULL * 1024 * 1024)
+#define OFD_PRECREATE_BATCH_SMALL 8
+
+/* Limit the returned fields marked valid to those that we actually might set */
+#define OFD_VALID_FLAGS (LA_TYPE | LA_MODE | LA_SIZE | LA_BLOCKS | \
+ LA_BLKSIZE | LA_ATIME | LA_MTIME | LA_CTIME)
+
+/* per-client-per-object persistent state (LRU) */
+struct ofd_mod_data {
+ cfs_list_t fmd_list; /* linked to fed_mod_list */
+ struct lu_fid fmd_fid; /* FID being written to */
+ __u64 fmd_mactime_xid; /* xid highest {m,a,c}time setattr */
+ cfs_time_t fmd_expire; /* time when the fmd should expire */
+ int fmd_refcount; /* reference counter - list holds 1 */
+};
+
+#define OFD_FMD_MAX_NUM_DEFAULT 128
+#define OFD_FMD_MAX_AGE_DEFAULT ((obd_timeout + 10) * HZ)
+
+enum {
+ LPROC_OFD_READ_BYTES = 0,
+ LPROC_OFD_WRITE_BYTES = 1,
+ LPROC_OFD_LAST,
+};
+
+/* for job stats */
+enum {
+ LPROC_OFD_STATS_READ = 0,
+ LPROC_OFD_STATS_WRITE = 1,
+ LPROC_OFD_STATS_SETATTR = 2,
+ LPROC_OFD_STATS_PUNCH = 3,
+ LPROC_OFD_STATS_SYNC = 4,
+ LPROC_OFD_STATS_LAST,
+};
+
+static inline void ofd_counter_incr(struct obd_export *exp, int opcode,
+ char *jobid, long amount)
+{
+ if (exp->exp_obd && exp->exp_obd->u.obt.obt_jobstats.ojs_hash &&
+ (exp_connect_flags(exp) & OBD_CONNECT_JOBSTATS))
+ lprocfs_job_stats_log(exp->exp_obd, jobid, opcode, amount);
+
+ if (exp->exp_nid_stats != NULL &&
+ exp->exp_nid_stats->nid_stats != NULL) {
+ if (opcode == LPROC_OFD_STATS_READ)
+ lprocfs_counter_add(exp->exp_nid_stats->nid_stats,
+ LPROC_OFD_READ_BYTES, amount);
+ else if (opcode == LPROC_OFD_STATS_WRITE)
+ lprocfs_counter_add(exp->exp_nid_stats->nid_stats,
+ LPROC_OFD_WRITE_BYTES, amount);
+ }
+}
+
+struct ofd_seq {
+ cfs_list_t os_list;
+ struct ost_id os_oi;
+ spinlock_t os_last_oid_lock;
+ struct mutex os_create_lock;
+ cfs_atomic_t os_refc;
+ struct dt_object *os_lastid_obj;
+ unsigned long os_destroys_in_progress:1;
+};
struct ofd_device {
- struct dt_device ofd_dt_dev;
+ struct dt_device ofd_dt_dev;
+ struct dt_device *ofd_osd;
+ struct obd_export *ofd_osd_exp;
+ struct dt_device_param ofd_dt_conf;
+ /* DLM name-space for meta-data locks maintained by this server */
+ struct ldlm_namespace *ofd_namespace;
+
+ /* transaction callbacks */
+ struct dt_txn_callback ofd_txn_cb;
+
+ /* last_rcvd file */
+ struct lu_target ofd_lut;
+ struct dt_object *ofd_health_check_file;
+
+ int ofd_subdir_count;
+
+ cfs_list_t ofd_seq_list;
+ rwlock_t ofd_seq_list_lock;
+ int ofd_seq_count;
+ int ofd_precreate_batch;
+ spinlock_t ofd_batch_lock;
+
+ /* protect all statfs-related counters */
+ spinlock_t ofd_osfs_lock;
+ /* statfs optimization: we cache a bit */
+ struct obd_statfs ofd_osfs;
+ __u64 ofd_osfs_age;
+ int ofd_blockbits;
+ /* writes which might be be accounted twice in ofd_osfs.os_bavail */
+ obd_size ofd_osfs_unstable;
+
+ /* counters used during statfs update, protected by ofd_osfs_lock.
+ * record when some statfs refresh are in progress */
+ int ofd_statfs_inflight;
+ /* track writes completed while statfs refresh is underway.
+ * tracking is only effective when ofd_statfs_inflight > 1 */
+ obd_size ofd_osfs_inflight;
+
+ /* grants: all values in bytes */
+ /* grant lock to protect all grant counters */
+ spinlock_t ofd_grant_lock;
+ /* total amount of dirty data reported by clients in incoming obdo */
+ obd_size ofd_tot_dirty;
+ /* sum of filesystem space granted to clients for async writes */
+ obd_size ofd_tot_granted;
+ /* grant used by I/Os in progress (between prepare and commit) */
+ obd_size ofd_tot_pending;
+ /* free space threshold over which we stop granting space to clients
+ * ofd_grant_ratio is stored as a fixed-point fraction using
+ * OFD_GRANT_RATIO_SHIFT of the remaining free space, not in percentage
+ * values */
+ int ofd_grant_ratio;
+ /* number of clients using grants */
+ int ofd_tot_granted_clients;
+
+ /* ofd mod data: ofd_device wide values */
+ int ofd_fmd_max_num; /* per ofd ofd_mod_data */
+ cfs_duration_t ofd_fmd_max_age; /* time to fmd expiry */
+
+ spinlock_t ofd_flags_lock;
+ unsigned long ofd_raid_degraded:1,
+ /* sync journal on writes */
+ ofd_syncjournal:1,
+ /* sync on lock cancel */
+ ofd_sync_lock_cancel:2,
+ /* shall we grant space to clients not
+ * supporting OBD_CONNECT_GRANT_PARAM? */
+ ofd_grant_compat_disable:1;
+ struct seq_server_site ofd_seq_site;
};
static inline struct ofd_device *ofd_dev(struct lu_device *d)
}
struct ofd_object {
- struct lu_object_header ofo_header;
+ struct lu_object_header ofo_header;
struct dt_object ofo_obj;
+ int ofo_ff_exists;
};
static inline struct ofd_object *ofd_obj(struct lu_object *o)
return container_of0(o, struct ofd_object, ofo_obj.do_lu);
}
+static inline int ofd_object_exists(struct ofd_object *obj)
+{
+ LASSERT(obj != NULL);
+ if (lu_object_is_dying(obj->ofo_obj.do_lu.lo_header))
+ return 0;
+ return lu_object_exists(&obj->ofo_obj.do_lu);
+}
+
+static inline struct dt_object *fo2dt(struct ofd_object *obj)
+{
+ return &obj->ofo_obj;
+}
+
+static inline struct dt_object *ofd_object_child(struct ofd_object *_obj)
+{
+ struct lu_object *lu = &(_obj)->ofo_obj.do_lu;
+
+ return container_of0(lu_object_next(lu), struct dt_object, do_lu);
+}
+
+static inline struct ofd_device *ofd_obj2dev(const struct ofd_object *fo)
+{
+ return ofd_dev(fo->ofo_obj.do_lu.lo_dev);
+}
+
+static inline struct lustre_capa *ofd_object_capa(const struct lu_env *env,
+ const struct ofd_object *obj)
+{
+ /* TODO: see mdd_object_capa() */
+ return BYPASS_CAPA;
+}
+
+static inline void ofd_read_lock(const struct lu_env *env,
+ struct ofd_object *fo)
+{
+ struct dt_object *next = ofd_object_child(fo);
+
+ next->do_ops->do_read_lock(env, next, 0);
+}
+
+static inline void ofd_read_unlock(const struct lu_env *env,
+ struct ofd_object *fo)
+{
+ struct dt_object *next = ofd_object_child(fo);
+
+ next->do_ops->do_read_unlock(env, next);
+}
+
+static inline void ofd_write_lock(const struct lu_env *env,
+ struct ofd_object *fo)
+{
+ struct dt_object *next = ofd_object_child(fo);
+
+ next->do_ops->do_write_lock(env, next, 0);
+}
+
+static inline void ofd_write_unlock(const struct lu_env *env,
+ struct ofd_object *fo)
+{
+ struct dt_object *next = ofd_object_child(fo);
+
+ next->do_ops->do_write_unlock(env, next);
+}
+
/*
* Common data shared by obdofd-level handlers. This is allocated per-thread
* to reduce stack consumption.
*/
struct ofd_thread_info {
- const struct lu_env *fti_env;
+ const struct lu_env *fti_env;
+
+ struct obd_export *fti_exp;
+ __u64 fti_xid;
+ __u64 fti_transno;
+ __u64 fti_pre_version;
+ __u32 fti_has_trans:1, /* has txn already */
+ fti_mult_trans:1;
+
+ struct lu_fid fti_fid;
+ struct lu_attr fti_attr;
+ struct lu_attr fti_attr2;
+ struct ldlm_res_id fti_resid;
+ struct filter_fid fti_mds_fid;
+ struct ost_id fti_ostid;
+ struct ofd_object *fti_obj;
+ union {
+ char name[64]; /* for ofd_init0() */
+ struct obd_statfs osfs; /* for obdofd_statfs() */
+ } fti_u;
+
+ /* Ops object filename */
+ struct lu_name fti_name;
+ struct dt_object_format fti_dof;
+ struct lu_buf fti_buf;
+ loff_t fti_off;
+
+ /* Space used by the I/O, used by grant code */
+ unsigned long fti_used;
+ struct ost_lvb fti_lvb;
};
+extern void target_recovery_fini(struct obd_device *obd);
+extern void target_recovery_init(struct lu_target *lut, svc_handler_t handler);
+
+/* ofd_capa.c */
+int ofd_update_capa_key(struct ofd_device *ofd, struct lustre_capa_key *key);
+int ofd_auth_capa(struct obd_export *exp, struct lu_fid *fid, obd_seq seq,
+ struct lustre_capa *capa, __u64 opc);
+void ofd_free_capa_keys(struct ofd_device *ofd);
+
/* ofd_dev.c */
extern struct lu_context_key ofd_thread_key;
+int ofd_postrecov(const struct lu_env *env, struct ofd_device *ofd);
/* ofd_obd.c */
extern struct obd_ops ofd_obd_ops;
+int ofd_statfs_internal(const struct lu_env *env, struct ofd_device *ofd,
+ struct obd_statfs *osfs, __u64 max_age,
+ int *from_cache);
+
+/* ofd_fs.c */
+obd_id ofd_seq_last_oid(struct ofd_seq *oseq);
+void ofd_seq_last_oid_set(struct ofd_seq *oseq, obd_id id);
+int ofd_seq_last_oid_write(const struct lu_env *env, struct ofd_device *ofd,
+ struct ofd_seq *oseq);
+int ofd_seqs_init(const struct lu_env *env, struct ofd_device *ofd);
+struct ofd_seq *ofd_seq_get(struct ofd_device *ofd, obd_seq seq);
+void ofd_seq_put(const struct lu_env *env, struct ofd_seq *oseq);
+
+int ofd_fs_setup(const struct lu_env *env, struct ofd_device *ofd,
+ struct obd_device *obd);
+void ofd_fs_cleanup(const struct lu_env *env, struct ofd_device *ofd);
+int ofd_precreate_batch(struct ofd_device *ofd, int batch);
+struct ofd_seq *ofd_seq_load(const struct lu_env *env, struct ofd_device *ofd,
+ obd_seq seq);
+void ofd_seqs_fini(const struct lu_env *env, struct ofd_device *ofd);
+
+/* ofd_io.c */
+int ofd_preprw(const struct lu_env *env,int cmd, struct obd_export *exp,
+ struct obdo *oa, int objcount, struct obd_ioobj *obj,
+ struct niobuf_remote *rnb, int *nr_local,
+ struct niobuf_local *lnb, struct obd_trans_info *oti,
+ struct lustre_capa *capa);
+int ofd_commitrw(const struct lu_env *env, int cmd, struct obd_export *exp,
+ struct obdo *oa, int objcount, struct obd_ioobj *obj,
+ struct niobuf_remote *rnb, int npages,
+ struct niobuf_local *lnb, struct obd_trans_info *oti,
+ int old_rc);
+
+/* ofd_trans.c */
+struct thandle *ofd_trans_create(const struct lu_env *env,
+ struct ofd_device *ofd);
+int ofd_trans_start(const struct lu_env *env,
+ struct ofd_device *ofd, struct ofd_object *fo,
+ struct thandle *th);
+void ofd_trans_stop(const struct lu_env *env, struct ofd_device *ofd,
+ struct thandle *th, int rc);
+int ofd_txn_stop_cb(const struct lu_env *env, struct thandle *txn,
+ void *cookie);
/* lproc_ofd.c */
+#ifdef LPROCFS
void lprocfs_ofd_init_vars(struct lprocfs_static_vars *lvars);
+void ofd_stats_counter_init(struct lprocfs_stats *stats);
+#else
+static void lprocfs_ofd_init_vars(struct lprocfs_static_vars *lvars)
+{
+ memset(lvars, 0, sizeof(*lvars));
+}
+static inline void ofd_stats_counter_init(struct lprocfs_stats *stats) {}
+#endif
+
+/* ofd_objects.c */
+struct ofd_object *ofd_object_find(const struct lu_env *env,
+ struct ofd_device *ofd,
+ const struct lu_fid *fid);
+struct ofd_object *ofd_object_find_or_create(const struct lu_env *env,
+ struct ofd_device *ofd,
+ const struct lu_fid *fid,
+ struct lu_attr *attr);
+int ofd_object_ff_check(const struct lu_env *env, struct ofd_object *fo);
+int ofd_precreate_objects(const struct lu_env *env, struct ofd_device *ofd,
+ obd_id id, struct ofd_seq *oseq, int nr, int sync);
+
+void ofd_object_put(const struct lu_env *env, struct ofd_object *fo);
+int ofd_attr_set(const struct lu_env *env, struct ofd_object *fo,
+ struct lu_attr *la, struct filter_fid *ff);
+int ofd_object_punch(const struct lu_env *env, struct ofd_object *fo,
+ __u64 start, __u64 end, struct lu_attr *la,
+ struct filter_fid *ff);
+int ofd_object_destroy(const struct lu_env *, struct ofd_object *, int);
+int ofd_attr_get(const struct lu_env *env, struct ofd_object *fo,
+ struct lu_attr *la);
+int ofd_attr_handle_ugid(const struct lu_env *env, struct ofd_object *fo,
+ struct lu_attr *la, int is_setattr);
+
+/* ofd_grants.c */
+#define OFD_GRANT_RATIO_SHIFT 8
+static inline __u64 ofd_grant_reserved(struct ofd_device *ofd, obd_size bavail)
+{
+ return (bavail * ofd->ofd_grant_ratio) >> OFD_GRANT_RATIO_SHIFT;
+}
+
+static inline int ofd_grant_ratio_conv(int percentage)
+{
+ return (percentage << OFD_GRANT_RATIO_SHIFT) / 100;
+}
+
+static inline int ofd_grant_param_supp(struct obd_export *exp)
+{
+ return !!(exp_connect_flags(exp) & OBD_CONNECT_GRANT_PARAM);
+}
+
+/* Blocksize used for client not supporting OBD_CONNECT_GRANT_PARAM.
+ * That's 4KB=2^12 which is the biggest block size known to work whatever
+ * the client's page size is. */
+#define COMPAT_BSIZE_SHIFT 12
+static inline int ofd_grant_compat(struct obd_export *exp,
+ struct ofd_device *ofd)
+{
+ /* Clients which don't support OBD_CONNECT_GRANT_PARAM cannot handle
+ * a block size > page size and consume PAGE_CACHE_SIZE of grant when
+ * dirtying a page regardless of the block size */
+ return !!(ofd_obd(ofd)->obd_self_export != exp &&
+ ofd->ofd_blockbits > COMPAT_BSIZE_SHIFT &&
+ !ofd_grant_param_supp(exp));
+}
+
+static inline int ofd_grant_prohibit(struct obd_export *exp,
+ struct ofd_device *ofd)
+{
+ /* When ofd_grant_compat_disable is set, we don't grant any space to
+ * clients not supporting OBD_CONNECT_GRANT_PARAM.
+ * Otherwise, space granted to such a client is inflated since it
+ * consumes PAGE_CACHE_SIZE of grant space per block */
+ return !!(ofd_grant_compat(exp, ofd) && ofd->ofd_grant_compat_disable);
+}
+
+void ofd_grant_sanity_check(struct obd_device *obd, const char *func);
+long ofd_grant_connect(const struct lu_env *env, struct obd_export *exp,
+ obd_size want, bool conservative);
+void ofd_grant_discard(struct obd_export *exp);
+void ofd_grant_prepare_read(const struct lu_env *env, struct obd_export *exp,
+ struct obdo *oa);
+void ofd_grant_prepare_write(const struct lu_env *env, struct obd_export *exp,
+ struct obdo *oa, struct niobuf_remote *rnb,
+ int niocount);
+void ofd_grant_commit(const struct lu_env *env, struct obd_export *exp, int rc);
+int ofd_grant_create(const struct lu_env *env, struct obd_export *exp, int *nr);
+
+/* ofd_fmd.c */
+int ofd_fmd_init(void);
+void ofd_fmd_exit(void);
+struct ofd_mod_data *ofd_fmd_find(struct obd_export *exp,
+ struct lu_fid *fid);
+struct ofd_mod_data *ofd_fmd_get(struct obd_export *exp,
+ struct lu_fid *fid);
+void ofd_fmd_put(struct obd_export *exp, struct ofd_mod_data *fmd);
+void ofd_fmd_expire(struct obd_export *exp);
+void ofd_fmd_cleanup(struct obd_export *exp);
+#ifdef DO_FMD_DROP
+void ofd_fmd_drop(struct obd_export *exp, struct lu_fid *fid);
+#else
+#define ofd_fmd_drop(exp, fid) do {} while (0)
+#endif
+
+/* ofd_dev.c */
+int ofd_fid_set_index(const struct lu_env *env, struct ofd_device *ofd,
+ int index);
+int ofd_fid_init(const struct lu_env *env, struct ofd_device *ofd);
+int ofd_fid_fini(const struct lu_env *env, struct ofd_device *ofd);
+
+/* ofd_lvb.c */
+extern struct ldlm_valblock_ops ofd_lvbo;
+
+/* ofd_dlm.c */
+int ofd_intent_policy(struct ldlm_namespace *ns, struct ldlm_lock **lockp,
+ void *req_cookie, ldlm_mode_t mode, __u64 flags,
+ void *data);
+
+static inline struct ofd_thread_info * ofd_info(const struct lu_env *env)
+{
+ struct ofd_thread_info *info;
+
+ info = lu_context_key_get(&env->le_ctx, &ofd_thread_key);
+ LASSERT(info);
+ LASSERT(info->fti_env);
+ LASSERT(info->fti_env == env);
+ return info;
+}
+
+static inline struct ofd_thread_info * ofd_info_init(const struct lu_env *env,
+ struct obd_export *exp)
+{
+ struct ofd_thread_info *info;
+
+ info = lu_context_key_get(&env->le_ctx, &ofd_thread_key);
+ LASSERT(info);
+ LASSERT(info->fti_exp == NULL);
+ LASSERT(info->fti_env == NULL);
+ LASSERT(info->fti_attr.la_valid == 0);
+
+ info->fti_env = env;
+ info->fti_exp = exp;
+ info->fti_pre_version = 0;
+ info->fti_transno = 0;
+ info->fti_has_trans = 0;
+ return info;
+}
+
+static inline void ofd_oti2info(struct ofd_thread_info *info,
+ struct obd_trans_info *oti)
+{
+ info->fti_xid = oti->oti_xid;
+ info->fti_transno = oti->oti_transno;
+ info->fti_pre_version = oti->oti_pre_version;
+}
+
+static inline void ofd_info2oti(struct ofd_thread_info *info,
+ struct obd_trans_info *oti)
+{
+ oti->oti_xid = info->fti_xid;
+ LASSERTF(ergo(oti->oti_transno > 0,
+ oti->oti_transno == info->fti_transno),
+ "Overwrite replay transno "LPX64" by "LPX64"\n",
+ oti->oti_transno, info->fti_transno);
+ oti->oti_transno = info->fti_transno;
+ oti->oti_pre_version = info->fti_pre_version;
+}
+
+/* sync on lock cancel is useless when we force a journal flush,
+ * and if we enable async journal commit, we should also turn on
+ * sync on lock cancel if it is not enabled already. */
+static inline void ofd_slc_set(struct ofd_device *ofd)
+{
+ if (ofd->ofd_syncjournal == 1)
+ ofd->ofd_sync_lock_cancel = NEVER_SYNC_ON_CANCEL;
+ else if (ofd->ofd_sync_lock_cancel == NEVER_SYNC_ON_CANCEL)
+ ofd->ofd_sync_lock_cancel = ALWAYS_SYNC_ON_CANCEL;
+}
+
+static inline void ofd_prepare_fidea(struct filter_fid *ff, struct obdo *oa)
+{
+ if (!(oa->o_valid & OBD_MD_FLGROUP))
+ ostid_set_seq_mdt0(&oa->o_oi);
+ /* packing fid and converting it to LE for storing into EA.
+ * Here ->o_stripe_idx should be filled by LOV and rest of
+ * fields - by client. */
+ ff->ff_parent.f_seq = cpu_to_le64(oa->o_parent_seq);
+ ff->ff_parent.f_oid = cpu_to_le32(oa->o_parent_oid);
+ /* XXX: we are ignoring o_parent_ver here, since this should
+ * be the same for all objects in this fileset. */
+ ff->ff_parent.f_ver = cpu_to_le32(oa->o_stripe_idx);
+}
+
+/* niobuf_remote has no rnb_ prefix in master */
+#define rnb_offset offset
+#define rnb_flags flags
+#define rnb_len len
+/* the same for niobuf_local */
+#define lnb_flags flags
+#define lnb_rc rc
#endif /* _OFD_INTERNAL_H */