* Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
* Use is subject to license terms.
*
- * Copyright (c) 2011, 2012, Whamcloud, Inc.
+ * Copyright (c) 2011, 2013, Intel Corporation.
*/
/*
* This file is part of Lustre, http://www.lustre.org/
* Author: Yury Umanets <umka@clusterfs.com>
*/
-#ifndef EXPORT_SYMTAB
-# define EXPORT_SYMTAB
-#endif
#define DEBUG_SUBSYSTEM S_FID
-#ifdef __KERNEL__
-# include <libcfs/libcfs.h>
-# include <linux/module.h>
-#else /* __KERNEL__ */
-# include <liblustre.h>
-#endif
-
-#include <obd.h>
-#include <obd_class.h>
+#include <libcfs/libcfs.h>
#include <dt_object.h>
#include <md_object.h>
#include <obd_support.h>
-#include <lustre_req_layout.h>
+#include <lustre_capa.h>
#include <lustre_fid.h>
+#include <lustre_fld.h>
#include "fid_internal.h"
-#ifdef __KERNEL__
-
static struct lu_buf *seq_store_buf(struct seq_thread_info *info)
{
struct lu_buf *buf;
};
void seq_update_cb(struct lu_env *env, struct thandle *th,
- struct dt_txn_commit_cb *cb, int err)
+ struct dt_txn_commit_cb *cb, int err)
{
- struct seq_update_callback *ccb;
- ccb = container_of0(cb, struct seq_update_callback, suc_cb);
- ccb->suc_seq->lss_need_sync = 0;
- cfs_list_del(&ccb->suc_cb.dcb_linkage);
- OBD_FREE_PTR(ccb);
-}
+ struct seq_update_callback *ccb;
-struct thandle *seq_store_trans_create(struct lu_server_seq *seq,
- const struct lu_env *env)
-{
- struct dt_device *dt_dev;
+ ccb = container_of0(cb, struct seq_update_callback, suc_cb);
- dt_dev = lu2dt_dev(seq->lss_obj->do_lu.lo_dev);
- return dt_trans_create(env, dt_dev);
-}
+ LASSERT(ccb->suc_seq != NULL);
-int seq_store_trans_start(struct lu_server_seq *seq, const struct lu_env *env,
- struct thandle *th)
-{
- struct dt_device *dt_dev;
- ENTRY;
-
- dt_dev = lu2dt_dev(seq->lss_obj->do_lu.lo_dev);
-
- return dt_trans_start(env, dt_dev, th);
+ ccb->suc_seq->lss_need_sync = 0;
+ OBD_FREE_PTR(ccb);
}
int seq_update_cb_add(struct thandle *th, struct lu_server_seq *seq)
{
- struct seq_update_callback *ccb;
- int rc;
- OBD_ALLOC_PTR(ccb);
- if (ccb == NULL)
- return -ENOMEM;
-
- ccb->suc_cb.dcb_func = seq_update_cb;
- CFS_INIT_LIST_HEAD(&ccb->suc_cb.dcb_linkage);
- ccb->suc_seq = seq;
- seq->lss_need_sync = 1;
- rc = dt_trans_cb_add(th, &ccb->suc_cb);
- if (rc)
- OBD_FREE_PTR(ccb);
- return rc;
-}
-
-int seq_declare_store_write(struct lu_server_seq *seq,
- const struct lu_env *env,
- struct thandle *th)
-{
- struct dt_object *dt_obj = seq->lss_obj;
- int rc;
- ENTRY;
-
- rc = dt_obj->do_body_ops->dbo_declare_write(env, dt_obj,
- sizeof(struct lu_seq_range),
- 0, th);
- return rc;
+ struct seq_update_callback *ccb;
+ struct dt_txn_commit_cb *dcb;
+ int rc;
+
+ OBD_ALLOC_PTR(ccb);
+ if (ccb == NULL)
+ return -ENOMEM;
+
+ ccb->suc_seq = seq;
+ seq->lss_need_sync = 1;
+
+ dcb = &ccb->suc_cb;
+ dcb->dcb_func = seq_update_cb;
+ CFS_INIT_LIST_HEAD(&dcb->dcb_linkage);
+ strncpy(dcb->dcb_name, "seq_update_cb", MAX_COMMIT_CB_STR_LEN);
+ dcb->dcb_name[MAX_COMMIT_CB_STR_LEN - 1] = '\0';
+
+ rc = dt_trans_cb_add(th, dcb);
+ if (rc)
+ OBD_FREE_PTR(ccb);
+ return rc;
}
/* This function implies that caller takes care about locking. */
-int seq_store_write(struct lu_server_seq *seq,
- const struct lu_env *env,
- struct thandle *th)
-{
- struct dt_object *dt_obj = seq->lss_obj;
- struct seq_thread_info *info;
- loff_t pos = 0;
- int rc;
- ENTRY;
-
- info = lu_context_key_get(&env->le_ctx, &seq_thread_key);
- LASSERT(info != NULL);
-
- /* Store ranges in le format. */
- range_cpu_to_le(&info->sti_space, &seq->lss_space);
-
- rc = dt_obj->do_body_ops->dbo_write(env, dt_obj,
- seq_store_buf(info),
- &pos, th, BYPASS_CAPA, 1);
- if (rc == sizeof(info->sti_space)) {
- CDEBUG(D_INFO, "%s: Space - "DRANGE"\n",
- seq->lss_name, PRANGE(&seq->lss_space));
- rc = 0;
- } else if (rc >= 0) {
- rc = -EIO;
- }
-
-
- RETURN(rc);
-}
-
int seq_store_update(const struct lu_env *env, struct lu_server_seq *seq,
struct lu_seq_range *out, int sync)
{
- struct dt_device *dt_dev;
- struct thandle *th;
- int rc;
- ENTRY;
-
- dt_dev = lu2dt_dev(seq->lss_obj->do_lu.lo_dev);
-
- th = seq_store_trans_create(seq, env);
- if (IS_ERR(th))
- RETURN(PTR_ERR(th));
-
- rc = seq_declare_store_write(seq, env, th);
- if (rc)
- GOTO(exit, rc);
-
- if (out != NULL) {
- rc = fld_declare_server_create(seq->lss_site->ms_server_fld,
- env, th);
- if (rc)
- GOTO(exit, rc);
- }
-
- rc = seq_store_trans_start(seq, env, th);
- if (rc)
- GOTO(exit, rc);
-
- rc = seq_store_write(seq, env, th);
- if (rc) {
- CERROR("%s: Can't write space data, rc %d\n",
- seq->lss_name, rc);
- GOTO(exit,rc);
- } else if (out != NULL) {
- rc = fld_server_create(seq->lss_site->ms_server_fld,
- env, out, th);
- if (rc) {
- CERROR("%s: Can't Update fld database, rc %d\n",
- seq->lss_name, rc);
- GOTO(exit,rc);
- }
- }
-
- /* next sequence update will need sync until this update is committed
- * in case of sync operation this is not needed obviously */
- if (!sync)
- /* if callback can't be added then sync always */
- sync = !!seq_update_cb_add(th, seq);
-
- th->th_sync |= sync;
+ struct dt_device *dt_dev = lu2dt_dev(seq->lss_obj->do_lu.lo_dev);
+ struct seq_thread_info *info;
+ struct thandle *th;
+ loff_t pos = 0;
+ int rc;
+
+ info = lu_context_key_get(&env->le_ctx, &seq_thread_key);
+ LASSERT(info != NULL);
+
+ th = dt_trans_create(env, dt_dev);
+ if (IS_ERR(th))
+ RETURN(PTR_ERR(th));
+
+ rc = dt_declare_record_write(env, seq->lss_obj,
+ sizeof(struct lu_seq_range), 0, th);
+ if (rc)
+ GOTO(exit, rc);
+
+ if (out != NULL) {
+ rc = fld_declare_server_create(env,
+ seq->lss_site->ss_server_fld,
+ out, th);
+ if (rc)
+ GOTO(exit, rc);
+ }
+
+ rc = dt_trans_start_local(env, dt_dev, th);
+ if (rc)
+ GOTO(exit, rc);
+
+ /* Store ranges in le format. */
+ range_cpu_to_le(&info->sti_space, &seq->lss_space);
+
+ rc = dt_record_write(env, seq->lss_obj, seq_store_buf(info), &pos, th);
+ if (rc) {
+ CERROR("%s: Can't write space data, rc %d\n",
+ seq->lss_name, rc);
+ GOTO(exit, rc);
+ } else if (out != NULL) {
+ rc = fld_server_create(env, seq->lss_site->ss_server_fld, out,
+ th);
+ if (rc) {
+ CERROR("%s: Can't Update fld database, rc %d\n",
+ seq->lss_name, rc);
+ GOTO(exit, rc);
+ }
+ }
+ /* next sequence update will need sync until this update is committed
+ * in case of sync operation this is not needed obviously */
+ if (!sync)
+ /* if callback can't be added then sync always */
+ sync = !!seq_update_cb_add(th, seq);
+
+ th->th_sync |= sync;
exit:
- dt_trans_stop(env, dt_dev, th);
- return rc;
+ dt_trans_stop(env, dt_dev, th);
+ return rc;
}
/*
int seq_store_read(struct lu_server_seq *seq,
const struct lu_env *env)
{
- struct dt_object *dt_obj = seq->lss_obj;
struct seq_thread_info *info;
loff_t pos = 0;
int rc;
info = lu_context_key_get(&env->le_ctx, &seq_thread_key);
LASSERT(info != NULL);
- rc = dt_obj->do_body_ops->dbo_read(env, dt_obj, seq_store_buf(info),
- &pos, BYPASS_CAPA);
+ rc = seq->lss_obj->do_body_ops->dbo_read(env, seq->lss_obj,
+ seq_store_buf(info),
+ &pos, BYPASS_CAPA);
if (rc == sizeof(info->sti_space)) {
range_le_to_cpu(&seq->lss_space, &info->sti_space);
rc = 0;
} else if (rc == 0) {
rc = -ENODATA;
- } else if (rc >= 0) {
+ } else if (rc > 0) {
CERROR("%s: Read only %d bytes of %d\n", seq->lss_name,
rc, (int)sizeof(info->sti_space));
rc = -EIO;
struct dt_device *dt)
{
struct dt_object *dt_obj;
- struct lu_fid fid;
+ struct lu_fid fid;
+ struct lu_attr attr;
+ struct dt_object_format dof;
const char *name;
int rc;
ENTRY;
- name = seq->lss_type == LUSTRE_SEQ_SERVER ?
- LUSTRE_SEQ_SRV_NAME : LUSTRE_SEQ_CTL_NAME;
+ name = seq->lss_type == LUSTRE_SEQ_SERVER ?
+ LUSTRE_SEQ_SRV_NAME : LUSTRE_SEQ_CTL_NAME;
+
+ if (seq->lss_type == LUSTRE_SEQ_SERVER)
+ lu_local_obj_fid(&fid, FID_SEQ_SRV_OID);
+ else
+ lu_local_obj_fid(&fid, FID_SEQ_CTL_OID);
+
+ memset(&attr, 0, sizeof(attr));
+ attr.la_valid = LA_MODE;
+ attr.la_mode = S_IFREG | 0666;
+ dof.dof_type = DFT_REGULAR;
- dt_obj = dt_store_open(env, dt, "", name, &fid);
+ dt_obj = dt_find_or_create(env, dt, &fid, &dof, &attr);
if (!IS_ERR(dt_obj)) {
seq->lss_obj = dt_obj;
rc = 0;
EXIT;
}
-#endif