if (IS_ERR(th))
RETURN(PTR_ERR(th));
+ tti_buf_lcd(tti);
rc = dt_declare_record_write(env, tgt->lut_last_rcvd,
- sizeof(struct lsd_client_data),
+ &tti->tti_buf,
ted->ted_lr_off, th);
if (rc)
GOTO(out, rc);
th->th_sync = sync;
+ tti_buf_lsd(tti);
rc = dt_declare_record_write(env, tgt->lut_last_rcvd,
- sizeof(struct lr_server_data),
- tti->tti_off, th);
+ &tti->tti_buf, tti->tti_off, th);
if (rc)
GOTO(out, rc);
struct lu_env env;
struct ptlrpc_request *req;
__u32 start_epoch;
- cfs_list_t client_list;
+ struct list_head client_list;
int rc;
if (tgt->lut_obd->obd_stopping)
* with resend requests. Move final list to separate one for processing
*/
spin_lock(&tgt->lut_obd->obd_recovery_task_lock);
- cfs_list_splice_init(&tgt->lut_obd->obd_final_req_queue, &client_list);
+ list_splice_init(&tgt->lut_obd->obd_final_req_queue, &client_list);
spin_unlock(&tgt->lut_obd->obd_recovery_task_lock);
/**
* go through list of exports participated in recovery and
* set new epoch for them
*/
- cfs_list_for_each_entry(req, &client_list, rq_list) {
+ list_for_each_entry(req, &client_list, rq_list) {
LASSERT(!req->rq_export->exp_delayed);
if (!req->rq_export->exp_vbr_failed)
tgt_client_epoch_update(&env, req->rq_export);
}
/** return list back at once */
spin_lock(&tgt->lut_obd->obd_recovery_task_lock);
- cfs_list_splice_init(&client_list, &tgt->lut_obd->obd_final_req_queue);
+ list_splice_init(&client_list, &tgt->lut_obd->obd_final_req_queue);
spin_unlock(&tgt->lut_obd->obd_recovery_task_lock);
/** update server epoch */
tgt_server_data_update(&env, tgt, 1);
dcb = &ccb->llcc_cb;
dcb->dcb_func = tgt_cb_last_committed;
CFS_INIT_LIST_HEAD(&dcb->dcb_linkage);
- strncpy(dcb->dcb_name, "tgt_cb_last_committed", MAX_COMMIT_CB_STR_LEN);
- dcb->dcb_name[MAX_COMMIT_CB_STR_LEN - 1] = '\0';
+ strlcpy(dcb->dcb_name, "tgt_cb_last_committed", sizeof(dcb->dcb_name));
rc = dt_trans_cb_add(th, dcb);
if (rc) {
ccb->lncc_exp->exp_client_uuid.uuid);
spin_lock(&ccb->lncc_exp->exp_lock);
- ccb->lncc_exp->exp_need_sync = 0;
+ /* XXX: Currently, we use per-export based sync/async policy for
+ * the update via OUT RPC, it is coarse-grained policy, and
+ * will be changed as per-request based by DNE II patches. */
+ if (!ccb->lncc_exp->exp_keep_sync)
+ ccb->lncc_exp->exp_need_sync = 0;
+
spin_unlock(&ccb->lncc_exp->exp_lock);
class_export_cb_put(ccb->lncc_exp);
dcb = &ccb->lncc_cb;
dcb->dcb_func = tgt_cb_new_client;
CFS_INIT_LIST_HEAD(&dcb->dcb_linkage);
- strncpy(dcb->dcb_name, "tgt_cb_new_client", MAX_COMMIT_CB_STR_LEN);
- dcb->dcb_name[MAX_COMMIT_CB_STR_LEN - 1] = '\0';
+ strlcpy(dcb->dcb_name, "tgt_cb_new_client", sizeof(dcb->dcb_name));
rc = dt_trans_cb_add(th, dcb);
if (rc) {
{
struct lu_target *tgt = cookie;
struct tgt_session_info *tsi;
+ struct tgt_thread_info *tti = tgt_th_info(env);
int rc;
/* if there is no session, then this transaction is not result of
if (tsi->tsi_exp == NULL)
return 0;
+ tti_buf_lcd(tti);
rc = dt_declare_record_write(env, tgt->lut_last_rcvd,
- sizeof(struct lsd_client_data),
+ &tti->tti_buf,
tsi->tsi_exp->exp_target_data.ted_lr_off,
th);
if (rc)
return rc;
+ tti_buf_lsd(tti);
rc = dt_declare_record_write(env, tgt->lut_last_rcvd,
- sizeof(struct lr_server_data), 0, th);
+ &tti->tti_buf, 0, th);
if (rc)
return rc;