* GPL HEADER END
*/
/*
- * Copyright (c) 2011, 2012, Intel Corporation.
+ * Copyright (c) 2013, 2017, Intel Corporation.
*/
/*
* lustre/target/tgt_handler.c
#define DEBUG_SUBSYSTEM S_CLASS
+#include <linux/user_namespace.h>
+#ifdef HAVE_UIDGID_HEADER
+# include <linux/uidgid.h>
+#endif
+
#include <obd.h>
#include <obd_class.h>
#include <obd_cksum.h>
+#include <lustre_lfsck.h>
+#include <lustre_nodemap.h>
+#include <lustre_acl.h>
#include "tgt_internal.h"
tsi->tsi_mdt_body = body;
- if (!(body->valid & OBD_MD_FLID))
+ if (!(body->mbo_valid & OBD_MD_FLID))
RETURN(0);
/* mdc_pack_body() doesn't check if fid is zero and set OBD_ML_FID
* in any case in pre-2.5 clients. Fix that here if needed */
- if (unlikely(fid_is_zero(&body->fid1)))
+ if (unlikely(fid_is_zero(&body->mbo_fid1)))
RETURN(0);
- if (!fid_is_sane(&body->fid1)) {
+ if (!fid_is_sane(&body->mbo_fid1)) {
CERROR("%s: invalid FID: "DFID"\n", tgt_name(tsi->tsi_tgt),
- PFID(&body->fid1));
+ PFID(&body->mbo_fid1));
RETURN(-EINVAL);
}
obj = lu_object_find(tsi->tsi_env,
&tsi->tsi_tgt->lut_bottom->dd_lu_dev,
- &body->fid1, NULL);
+ &body->mbo_fid1, NULL);
if (!IS_ERR(obj)) {
if ((flags & HABEO_CORPUS) && !lu_object_exists(obj)) {
lu_object_put(tsi->tsi_env, obj);
- /* for capability renew ENOENT will be handled in
- * mdt_renew_capa */
- if (body->valid & OBD_MD_FLOSSCAPA)
- rc = 0;
- else
- rc = -ENOENT;
+ rc = -ENOENT;
} else {
tsi->tsi_corpus = obj;
rc = 0;
} else {
rc = PTR_ERR(obj);
}
+
+ tsi->tsi_fid = body->mbo_fid1;
+
RETURN(rc);
}
*/
int tgt_validate_obdo(struct tgt_session_info *tsi, struct obdo *oa)
{
- int rc;
-
+ struct ost_id *oi = &oa->o_oi;
+ u64 seq = ostid_seq(oi);
+ u64 id = ostid_id(oi);
+ int rc;
ENTRY;
if (unlikely(!(exp_connect_flags(tsi->tsi_exp) & OBD_CONNECT_FID) &&
- fid_seq_is_echo(oa->o_oi.oi.oi_seq))) {
+ fid_seq_is_echo(seq))) {
/* Sigh 2.[123] client still sends echo req with oi_id = 0
* during create, and we will reset this to 1, since this
* oi_id is basically useless in the following create process,
* but oi_id == 0 will make it difficult to tell whether it is
* real FID or ost_id. */
- oa->o_oi.oi_fid.f_oid = oa->o_oi.oi.oi_id ?: 1;
- oa->o_oi.oi_fid.f_seq = FID_SEQ_ECHO;
- oa->o_oi.oi_fid.f_ver = 0;
+ oi->oi_fid.f_seq = FID_SEQ_ECHO;
+ oi->oi_fid.f_oid = id ?: 1;
+ oi->oi_fid.f_ver = 0;
} else {
- if (unlikely((oa->o_valid & OBD_MD_FLID &&
- ostid_id(&oa->o_oi) == 0)))
+ struct tgt_thread_info *tti = tgt_th_info(tsi->tsi_env);
+
+ if (unlikely((oa->o_valid & OBD_MD_FLID) && id == 0))
GOTO(out, rc = -EPROTO);
/* Note: this check might be forced in 2.5 or 2.6, i.e.
* all of the requests are required to setup FLGROUP */
if (unlikely(!(oa->o_valid & OBD_MD_FLGROUP))) {
- ostid_set_seq_mdt0(&oa->o_oi);
+ ostid_set_seq_mdt0(oi);
oa->o_valid |= OBD_MD_FLGROUP;
+ seq = ostid_seq(oi);
}
- if (unlikely(!(fid_seq_is_idif(ostid_seq(&oa->o_oi)) ||
- fid_seq_is_mdt0(ostid_seq(&oa->o_oi)) ||
- fid_seq_is_norm(ostid_seq(&oa->o_oi)) ||
- fid_seq_is_echo(ostid_seq(&oa->o_oi)))))
+ if (unlikely(!(fid_seq_is_idif(seq) || fid_seq_is_mdt0(seq) ||
+ fid_seq_is_norm(seq) || fid_seq_is_echo(seq))))
GOTO(out, rc = -EPROTO);
+
+ rc = ostid_to_fid(&tti->tti_fid1, oi,
+ tsi->tsi_tgt->lut_lsd.lsd_osd_index);
+ if (unlikely(rc != 0))
+ GOTO(out, rc);
+
+ oi->oi_fid = tti->tti_fid1;
}
+
RETURN(0);
+
out:
CERROR("%s: client %s sent bad object "DOSTID": rc = %d\n",
tgt_name(tsi->tsi_tgt), obd_export_nid2str(tsi->tsi_exp),
- ostid_seq(&oa->o_oi), ostid_id(&oa->o_oi), rc);
+ seq, id, rc);
return rc;
}
EXPORT_SYMBOL(tgt_validate_obdo);
+static int tgt_io_data_unpack(struct tgt_session_info *tsi, struct ost_id *oi)
+{
+ unsigned max_brw;
+ struct niobuf_remote *rnb;
+ struct obd_ioobj *ioo;
+ int obj_count;
+
+ ENTRY;
+
+ ioo = req_capsule_client_get(tsi->tsi_pill, &RMF_OBD_IOOBJ);
+ if (ioo == NULL)
+ RETURN(-EPROTO);
+
+ rnb = req_capsule_client_get(tsi->tsi_pill, &RMF_NIOBUF_REMOTE);
+ if (rnb == NULL)
+ RETURN(-EPROTO);
+
+ max_brw = ioobj_max_brw_get(ioo);
+ if (unlikely((max_brw & (max_brw - 1)) != 0)) {
+ CERROR("%s: client %s sent bad ioobj max %u for "DOSTID
+ ": rc = %d\n", tgt_name(tsi->tsi_tgt),
+ obd_export_nid2str(tsi->tsi_exp), max_brw,
+ POSTID(oi), -EPROTO);
+ RETURN(-EPROTO);
+ }
+ ioo->ioo_oid = *oi;
+
+ obj_count = req_capsule_get_size(tsi->tsi_pill, &RMF_OBD_IOOBJ,
+ RCL_CLIENT) / sizeof(*ioo);
+ if (obj_count == 0) {
+ CERROR("%s: short ioobj\n", tgt_name(tsi->tsi_tgt));
+ RETURN(-EPROTO);
+ } else if (obj_count > 1) {
+ CERROR("%s: too many ioobjs (%d)\n", tgt_name(tsi->tsi_tgt),
+ obj_count);
+ RETURN(-EPROTO);
+ }
+
+ if (ioo->ioo_bufcnt == 0) {
+ CERROR("%s: ioo has zero bufcnt\n", tgt_name(tsi->tsi_tgt));
+ RETURN(-EPROTO);
+ }
+
+ if (ioo->ioo_bufcnt > PTLRPC_MAX_BRW_PAGES) {
+ DEBUG_REQ(D_RPCTRACE, tgt_ses_req(tsi),
+ "bulk has too many pages (%d)",
+ ioo->ioo_bufcnt);
+ RETURN(-EPROTO);
+ }
+
+ RETURN(0);
+}
+
static int tgt_ost_body_unpack(struct tgt_session_info *tsi, __u32 flags)
{
struct ost_body *body;
struct req_capsule *pill = tsi->tsi_pill;
- struct lustre_capa *capa;
- struct obd_ioobj *ioo;
+ struct lu_nodemap *nodemap;
int rc;
ENTRY;
if (rc)
RETURN(rc);
- if (body->oa.o_valid & OBD_MD_FLOSSCAPA) {
- capa = req_capsule_client_get(tsi->tsi_pill, &RMF_CAPA1);
- if (capa == NULL) {
- CERROR("%s: OSSCAPA flag is set without capability\n",
- tgt_name(tsi->tsi_tgt));
- RETURN(-EFAULT);
- }
- }
+ nodemap = nodemap_get_from_exp(tsi->tsi_exp);
+ if (IS_ERR(nodemap))
+ RETURN(PTR_ERR(nodemap));
+
+ body->oa.o_uid = nodemap_map_id(nodemap, NODEMAP_UID,
+ NODEMAP_CLIENT_TO_FS,
+ body->oa.o_uid);
+ body->oa.o_gid = nodemap_map_id(nodemap, NODEMAP_GID,
+ NODEMAP_CLIENT_TO_FS,
+ body->oa.o_gid);
+ nodemap_putref(nodemap);
tsi->tsi_ost_body = body;
+ tsi->tsi_fid = body->oa.o_oi.oi_fid;
if (req_capsule_has_field(pill, &RMF_OBD_IOOBJ, RCL_CLIENT)) {
- unsigned max_brw;
- struct niobuf_remote *rnb;
-
- ioo = req_capsule_client_get(pill, &RMF_OBD_IOOBJ);
- if (ioo == NULL)
- RETURN(-EPROTO);
-
- rnb = req_capsule_client_get(pill, &RMF_NIOBUF_REMOTE);
- if (rnb == NULL)
- RETURN(-EPROTO);
-
- max_brw = ioobj_max_brw_get(ioo);
- if (unlikely((max_brw & (max_brw - 1)) != 0)) {
- CERROR("%s: client %s sent bad ioobj max %u for "DOSTID
- ": rc = %d\n", tgt_name(tsi->tsi_tgt),
- obd_export_nid2str(tsi->tsi_exp), max_brw,
- POSTID(&body->oa.o_oi), -EPROTO);
- RETURN(-EPROTO);
- }
- ioo->ioo_oid = body->oa.o_oi;
+ rc = tgt_io_data_unpack(tsi, &body->oa.o_oi);
+ if (rc < 0)
+ RETURN(rc);
}
if (!(body->oa.o_valid & OBD_MD_FLID)) {
}
}
- rc = ostid_to_fid(&tsi->tsi_fid, &body->oa.o_oi, 0);
- if (rc != 0)
- RETURN(rc);
-
- if (!fid_is_sane(&tsi->tsi_fid)) {
- CERROR("%s: invalid FID: "DFID"\n", tgt_name(tsi->tsi_tgt),
- PFID(&tsi->tsi_fid));
- RETURN(-EINVAL);
- }
-
ost_fid_build_resid(&tsi->tsi_fid, &tsi->tsi_resid);
/*
RETURN(rc);
}
-static int tgt_unpack_req_pack_rep(struct tgt_session_info *tsi, __u32 flags)
+/*
+ * Do necessary preprocessing according to handler ->th_flags.
+ */
+static int tgt_request_preprocess(struct tgt_session_info *tsi,
+ struct tgt_handler *h,
+ struct ptlrpc_request *req)
{
struct req_capsule *pill = tsi->tsi_pill;
- int rc;
+ __u32 flags = h->th_flags;
+ int rc = 0;
ENTRY;
- if (req_capsule_has_field(pill, &RMF_MDT_BODY, RCL_CLIENT)) {
- rc = tgt_mdt_body_unpack(tsi, flags);
- } else if (req_capsule_has_field(pill, &RMF_OST_BODY, RCL_CLIENT)) {
- rc = tgt_ost_body_unpack(tsi, flags);
- } else {
- rc = 0;
+ if (tsi->tsi_preprocessed)
+ RETURN(0);
+
+ LASSERT(h->th_act != NULL);
+ LASSERT(h->th_opc == lustre_msg_get_opc(req->rq_reqmsg));
+ LASSERT(current->journal_info == NULL);
+
+ LASSERT(ergo(flags & (HABEO_CORPUS | HABEO_REFERO),
+ h->th_fmt != NULL));
+ if (h->th_fmt != NULL) {
+ req_capsule_set(pill, h->th_fmt);
+ if (req_capsule_has_field(pill, &RMF_MDT_BODY, RCL_CLIENT)) {
+ rc = tgt_mdt_body_unpack(tsi, flags);
+ if (rc < 0)
+ RETURN(rc);
+ } else if (req_capsule_has_field(pill, &RMF_OST_BODY,
+ RCL_CLIENT)) {
+ rc = tgt_ost_body_unpack(tsi, flags);
+ if (rc < 0)
+ RETURN(rc);
+ }
}
- if (rc == 0 && flags & HABEO_REFERO) {
- /* Pack reply */
- if (req_capsule_has_field(pill, &RMF_MDT_MD, RCL_SERVER))
- req_capsule_set_size(pill, &RMF_MDT_MD, RCL_SERVER,
- tsi->tsi_mdt_body->eadatasize);
- if (req_capsule_has_field(pill, &RMF_LOGCOOKIES, RCL_SERVER))
- req_capsule_set_size(pill, &RMF_LOGCOOKIES,
- RCL_SERVER, 0);
+ if (flags & MUTABOR && tgt_conn_flags(tsi) & OBD_CONNECT_RDONLY)
+ RETURN(-EROFS);
+
+ if (flags & HABEO_CLAVIS) {
+ struct ldlm_request *dlm_req;
- rc = req_capsule_server_pack(pill);
+ LASSERT(h->th_fmt != NULL);
+
+ dlm_req = req_capsule_client_get(pill, &RMF_DLM_REQ);
+ if (dlm_req != NULL) {
+ union ldlm_wire_policy_data *policy =
+ &dlm_req->lock_desc.l_policy_data;
+
+ if (unlikely(dlm_req->lock_desc.l_resource.lr_type ==
+ LDLM_IBITS &&
+ (policy->l_inodebits.bits |
+ policy->l_inodebits.try_bits) == 0)) {
+ /*
+ * Lock without inodebits makes no sense and
+ * will oops later in ldlm. If client miss to
+ * set such bits, do not trigger ASSERTION.
+ *
+ * For liblustre flock case, it maybe zero.
+ */
+ rc = -EPROTO;
+ } else {
+ tsi->tsi_dlm_req = dlm_req;
+ }
+ } else {
+ rc = -EFAULT;
+ }
}
+ tsi->tsi_preprocessed = 1;
RETURN(rc);
}
{
int serious = 0;
int rc;
- __u32 flags;
+ __u32 opc = lustre_msg_get_opc(req->rq_reqmsg);
ENTRY;
- LASSERT(h->th_act != NULL);
- LASSERT(h->th_opc == lustre_msg_get_opc(req->rq_reqmsg));
- LASSERT(current->journal_info == NULL);
+
+ /* When dealing with sec context requests, no export is associated yet,
+ * because these requests are sent before *_CONNECT requests.
+ * A NULL req->rq_export means the normal *_common_slice handlers will
+ * not be called, because there is no reference to the target.
+ * So deal with them by hand and jump directly to target_send_reply().
+ */
+ switch (opc) {
+ case SEC_CTX_INIT:
+ case SEC_CTX_INIT_CONT:
+ case SEC_CTX_FINI:
+ CFS_FAIL_TIMEOUT(OBD_FAIL_SEC_CTX_HDL_PAUSE, cfs_fail_val);
+ GOTO(out, rc = 0);
+ }
/*
* Checking for various OBD_FAIL_$PREF_$OPC_NET codes. _Do_ not try
*/
if (OBD_FAIL_CHECK_ORSET(h->th_fail_id, OBD_FAIL_ONCE))
RETURN(0);
+ if (unlikely(lustre_msg_get_opc(req->rq_reqmsg) == MDS_REINT &&
+ OBD_FAIL_CHECK(OBD_FAIL_MDS_REINT_MULTI_NET)))
+ RETURN(0);
- rc = 0;
- flags = h->th_flags;
- LASSERT(ergo(flags & (HABEO_CORPUS | HABEO_REFERO),
- h->th_fmt != NULL));
- if (h->th_fmt != NULL) {
- req_capsule_set(tsi->tsi_pill, h->th_fmt);
- rc = tgt_unpack_req_pack_rep(tsi, flags);
- }
-
- if (rc == 0 && flags & MUTABOR &&
- tgt_conn_flags(tsi) & OBD_CONNECT_RDONLY)
- rc = -EROFS;
-
- if (rc == 0 && flags & HABEO_CLAVIS) {
- struct ldlm_request *dlm_req;
-
- LASSERT(h->th_fmt != NULL);
-
- dlm_req = req_capsule_client_get(tsi->tsi_pill, &RMF_DLM_REQ);
- if (dlm_req != NULL) {
- if (unlikely(dlm_req->lock_desc.l_resource.lr_type ==
- LDLM_IBITS &&
- dlm_req->lock_desc.l_policy_data.\
- l_inodebits.bits == 0)) {
- /*
- * Lock without inodebits makes no sense and
- * will oops later in ldlm. If client miss to
- * set such bits, do not trigger ASSERTION.
- *
- * For liblustre flock case, it maybe zero.
- */
- rc = -EPROTO;
- } else {
- tsi->tsi_dlm_req = dlm_req;
- }
- } else {
- rc = -EFAULT;
+ rc = tgt_request_preprocess(tsi, h, req);
+ /* pack reply if reply format is fixed */
+ if (rc == 0 && h->th_flags & HABEO_REFERO) {
+ /* Pack reply */
+ if (req_capsule_has_field(tsi->tsi_pill, &RMF_MDT_MD,
+ RCL_SERVER))
+ req_capsule_set_size(tsi->tsi_pill, &RMF_MDT_MD,
+ RCL_SERVER,
+ tsi->tsi_mdt_body->mbo_eadatasize);
+ if (req_capsule_has_field(tsi->tsi_pill, &RMF_LOGCOOKIES,
+ RCL_SERVER))
+ req_capsule_set_size(tsi->tsi_pill, &RMF_LOGCOOKIES,
+ RCL_SERVER, 0);
+ if (req_capsule_has_field(tsi->tsi_pill, &RMF_ACL, RCL_SERVER))
+ req_capsule_set_size(tsi->tsi_pill,
+ &RMF_ACL, RCL_SERVER,
+ LUSTRE_POSIX_ACL_MAX_SIZE_OLD);
+
+ if (req_capsule_has_field(tsi->tsi_pill, &RMF_SHORT_IO,
+ RCL_SERVER)) {
+ struct niobuf_remote *remote_nb =
+ req_capsule_client_get(tsi->tsi_pill,
+ &RMF_NIOBUF_REMOTE);
+ struct ost_body *body = tsi->tsi_ost_body;
+
+ req_capsule_set_size(tsi->tsi_pill, &RMF_SHORT_IO,
+ RCL_SERVER,
+ (body->oa.o_flags & OBD_FL_SHORT_IO) ?
+ remote_nb[0].rnb_len : 0);
}
+
+ rc = req_capsule_server_pack(tsi->tsi_pill);
}
if (likely(rc == 0)) {
if (likely(rc == 0 && req->rq_export))
target_committed_to_req(req);
+out:
target_send_reply(req, rc, tsi->tsi_reply_fail_id);
RETURN(0);
}
*process = 1;
RETURN(0);
case MDS_CLOSE:
- case MDS_DONE_WRITING:
case MDS_SYNC: /* used in unmounting */
case OBD_PING:
case MDS_REINT:
- case UPDATE_OBJ:
+ case OUT_UPDATE:
case SEQ_QUERY:
case FLD_QUERY:
+ case FLD_READ:
case LDLM_ENQUEUE:
case OST_CREATE:
case OST_DESTROY:
case OST_SETATTR:
case OST_SYNC:
case OST_WRITE:
+ case MDS_HSM_PROGRESS:
+ case MDS_HSM_STATE_SET:
+ case MDS_HSM_REQUEST:
*process = target_queue_recovery_request(req, obd);
RETURN(0);
* -ve: abort immediately with the given error code;
* 0: send reply with error code in req->rq_status;
*/
-int tgt_handle_recovery(struct ptlrpc_request *req, int reply_fail_id)
+static int tgt_handle_recovery(struct ptlrpc_request *req, int reply_fail_id)
{
ENTRY;
/* sanity check: if the xid matches, the request must be marked as a
* resent or replayed */
- if (req_xid_is_last(req)) {
+ if (req_can_reconstruct(req, NULL)) {
if (!(lustre_msg_get_flags(req->rq_reqmsg) &
(MSG_RESENT | MSG_REPLAY))) {
- DEBUG_REQ(D_WARNING, req, "rq_xid "LPU64" matches "
- "last_xid, expected REPLAY or RESENT flag "
+ DEBUG_REQ(D_WARNING, req, "rq_xid %llu matches "
+ "saved xid, expected REPLAY or RESENT flag "
"(%x)", req->rq_xid,
lustre_msg_get_flags(req->rq_reqmsg));
req->rq_status = -ENOTCONN;
RETURN(+1);
}
+/* Initial check for request, it is validation mostly */
+static struct tgt_handler *tgt_handler_find_check(struct ptlrpc_request *req)
+{
+ struct tgt_handler *h;
+ struct tgt_opc_slice *s;
+ struct lu_target *tgt;
+ __u32 opc = lustre_msg_get_opc(req->rq_reqmsg);
+
+ ENTRY;
+
+ tgt = class_exp2tgt(req->rq_export);
+ if (unlikely(tgt == NULL)) {
+ DEBUG_REQ(D_ERROR, req, "%s: No target for connected export\n",
+ class_exp2obd(req->rq_export)->obd_name);
+ RETURN(ERR_PTR(-EINVAL));
+ }
+
+ for (s = tgt->lut_slice; s->tos_hs != NULL; s++)
+ if (s->tos_opc_start <= opc && opc < s->tos_opc_end)
+ break;
+
+ /* opcode was not found in slice */
+ if (unlikely(s->tos_hs == NULL)) {
+ CERROR("%s: no handlers for opcode 0x%x\n", tgt_name(tgt),
+ opc);
+ RETURN(ERR_PTR(-ENOTSUPP));
+ }
+
+ LASSERT(opc >= s->tos_opc_start && opc < s->tos_opc_end);
+ h = s->tos_hs + (opc - s->tos_opc_start);
+ if (unlikely(h->th_opc == 0)) {
+ CERROR("%s: unsupported opcode 0x%x\n", tgt_name(tgt), opc);
+ RETURN(ERR_PTR(-ENOTSUPP));
+ }
+
+ RETURN(h);
+}
+
+static int process_req_last_xid(struct ptlrpc_request *req)
+{
+ __u64 last_xid;
+ ENTRY;
+
+ /* check request's xid is consistent with export's last_xid */
+ last_xid = lustre_msg_get_last_xid(req->rq_reqmsg);
+ if (last_xid > req->rq_export->exp_last_xid)
+ req->rq_export->exp_last_xid = last_xid;
+
+ if (req->rq_xid == 0 ||
+ (req->rq_xid <= req->rq_export->exp_last_xid)) {
+ DEBUG_REQ(D_ERROR, req, "Unexpected xid %llx vs. "
+ "last_xid %llx\n", req->rq_xid,
+ req->rq_export->exp_last_xid);
+ /* Some request is allowed to be sent during replay,
+ * such as OUT update requests, FLD requests, so it
+ * is possible that replay requests has smaller XID
+ * than the exp_last_xid.
+ *
+ * Some non-replay requests may have smaller XID as
+ * well:
+ *
+ * - Client send a no_resend RPC, like statfs;
+ * - The RPC timedout (or some other error) on client,
+ * then it's removed from the unreplied list;
+ * - Client send some other request to bump the
+ * exp_last_xid on server;
+ * - The former RPC got chance to be processed;
+ */
+ if (!(lustre_msg_get_flags(req->rq_reqmsg) & MSG_REPLAY))
+ RETURN(-EPROTO);
+ }
+
+ /* The "last_xid" is the minimum xid among unreplied requests,
+ * if the request is from the previous connection, its xid can
+ * still be larger than "exp_last_xid", then the above check of
+ * xid is not enough to determine whether the request is delayed.
+ *
+ * For example, if some replay request was delayed and caused
+ * timeout at client and the replay is restarted, the delayed
+ * replay request will have the larger xid than "exp_last_xid"
+ */
+ if (req->rq_export->exp_conn_cnt >
+ lustre_msg_get_conn_cnt(req->rq_reqmsg))
+ RETURN(-ESTALE);
+
+ /* try to release in-memory reply data */
+ if (tgt_is_multimodrpcs_client(req->rq_export)) {
+ tgt_handle_received_xid(req->rq_export,
+ lustre_msg_get_last_xid(req->rq_reqmsg));
+ if (!(lustre_msg_get_flags(req->rq_reqmsg) &
+ (MSG_RESENT | MSG_REPLAY)))
+ tgt_handle_tag(req->rq_export,
+ lustre_msg_get_tag(req->rq_reqmsg));
+ }
+ RETURN(0);
+}
+
int tgt_request_handle(struct ptlrpc_request *req)
{
struct tgt_session_info *tsi = tgt_ses_info(req->rq_svc_thread->t_env);
struct lustre_msg *msg = req->rq_reqmsg;
struct tgt_handler *h;
- struct tgt_opc_slice *s;
struct lu_target *tgt;
int request_fail_id = 0;
__u32 opc = lustre_msg_get_opc(msg);
+ struct obd_device *obd;
int rc;
-
+ bool is_connect = false;
ENTRY;
+ if (unlikely(OBD_FAIL_CHECK(OBD_FAIL_TGT_RECOVERY_REQ_RACE))) {
+ if (cfs_fail_val == 0 &&
+ lustre_msg_get_opc(msg) != OBD_PING &&
+ lustre_msg_get_flags(msg) & MSG_REQ_REPLAY_DONE) {
+ struct l_wait_info lwi = { 0 };
+
+ cfs_fail_val = 1;
+ cfs_race_state = 0;
+ l_wait_event(cfs_race_waitq, (cfs_race_state == 1),
+ &lwi);
+ }
+ }
+
/* Refill the context, to make sure all thread keys are allocated */
lu_env_refill(req->rq_svc_thread->t_env);
* target, otherwise that should be connect operation */
if (opc == MDS_CONNECT || opc == OST_CONNECT ||
opc == MGS_CONNECT) {
+ is_connect = true;
req_capsule_set(&req->rq_pill, &RQF_CONNECT);
rc = target_handle_connect(req);
if (rc != 0) {
}
if (unlikely(!class_connected_export(req->rq_export))) {
+ if (opc == SEC_CTX_INIT || opc == SEC_CTX_INIT_CONT ||
+ opc == SEC_CTX_FINI) {
+ /* sec context initialization has to be handled
+ * by hand in tgt_handle_request0() */
+ tsi->tsi_reply_fail_id = OBD_FAIL_SEC_CTX_INIT_NET;
+ h = NULL;
+ GOTO(handle_recov, rc = 0);
+ }
CDEBUG(D_HA, "operation %d on unconnected OST from %s\n",
opc, libcfs_id2str(req->rq_peer));
req->rq_status = -ENOTCONN;
else
tsi->tsi_jobid = NULL;
- request_fail_id = tgt->lut_request_fail_id;
- tsi->tsi_reply_fail_id = tgt->lut_reply_fail_id;
-
- for (s = tgt->lut_slice; s->tos_hs != NULL; s++)
- if (s->tos_opc_start <= opc && opc < s->tos_opc_end)
- break;
-
- /* opcode was not found in slice */
- if (unlikely(s->tos_hs == NULL)) {
- CERROR("%s: no handlers for opcode 0x%x\n", tgt_name(tgt), opc);
- req->rq_status = -ENOTSUPP;
+ if (tgt == NULL) {
+ DEBUG_REQ(D_ERROR, req, "%s: No target for connected export\n",
+ class_exp2obd(req->rq_export)->obd_name);
+ req->rq_status = -EINVAL;
rc = ptlrpc_error(req);
GOTO(out, rc);
}
- if (CFS_FAIL_CHECK_ORSET(request_fail_id, CFS_FAIL_ONCE))
- GOTO(out, rc = 0);
+ /* Skip last_xid processing for the recovery thread, otherwise, the
+ * last_xid on same request could be processed twice: first time when
+ * processing the incoming request, second time when the request is
+ * being processed by recovery thread. */
+ obd = class_exp2obd(req->rq_export);
+ if (is_connect) {
+ /* reset the exp_last_xid on each connection. */
+ req->rq_export->exp_last_xid = 0;
+ } else if (obd->obd_recovery_data.trd_processing_task !=
+ current_pid()) {
+ rc = process_req_last_xid(req);
+ if (rc) {
+ req->rq_status = rc;
+ rc = ptlrpc_error(req);
+ GOTO(out, rc);
+ }
+ }
- LASSERT(current->journal_info == NULL);
+ request_fail_id = tgt->lut_request_fail_id;
+ tsi->tsi_reply_fail_id = tgt->lut_reply_fail_id;
- LASSERT(opc >= s->tos_opc_start && opc < s->tos_opc_end);
- h = s->tos_hs + (opc - s->tos_opc_start);
- if (unlikely(h->th_opc == 0)) {
- CERROR("%s: unsupported opcode 0x%x\n", tgt_name(tgt), opc);
- req->rq_status = -ENOTSUPP;
+ h = tgt_handler_find_check(req);
+ if (IS_ERR(h)) {
+ req->rq_status = PTR_ERR(h);
rc = ptlrpc_error(req);
GOTO(out, rc);
}
+ LASSERTF(h->th_opc == opc, "opcode mismatch %d != %d\n",
+ h->th_opc, opc);
+
+ if (CFS_FAIL_CHECK_ORSET(request_fail_id, CFS_FAIL_ONCE))
+ GOTO(out, rc = 0);
+
rc = lustre_msg_check_version(msg, h->th_version);
if (unlikely(rc)) {
DEBUG_REQ(D_ERROR, req, "%s: drop mal-formed request, version"
GOTO(out, rc);
}
+handle_recov:
rc = tgt_handle_recovery(req, tsi->tsi_reply_fail_id);
if (likely(rc == 1)) {
- LASSERTF(h->th_opc == opc, "opcode mismatch %d != %d\n",
- h->th_opc, opc);
rc = tgt_handle_request0(tsi, h, req);
if (rc)
GOTO(out, rc);
EXIT;
out:
req_capsule_fini(tsi->tsi_pill);
- tsi->tsi_pill = NULL;
if (tsi->tsi_corpus != NULL) {
lu_object_put(tsi->tsi_env, tsi->tsi_corpus);
tsi->tsi_corpus = NULL;
}
- tsi->tsi_env = NULL;
- tsi->tsi_mdt_body = NULL;
- tsi->tsi_dlm_req = NULL;
- fid_zero(&tsi->tsi_fid);
- memset(&tsi->tsi_resid, 0, sizeof tsi->tsi_resid);
return rc;
}
EXPORT_SYMBOL(tgt_request_handle);
-void tgt_counter_incr(struct obd_export *exp, int opcode)
-{
- lprocfs_counter_incr(exp->exp_obd->obd_stats, opcode);
- if (exp->exp_nid_stats && exp->exp_nid_stats->nid_stats != NULL)
- lprocfs_counter_incr(exp->exp_nid_stats->nid_stats, opcode);
-}
-EXPORT_SYMBOL(tgt_counter_incr);
-
-/*
- * Unified target generic handlers.
- */
-
-/*
- * Security functions
- */
-static inline void tgt_init_sec_none(struct obd_connect_data *reply)
-{
- reply->ocd_connect_flags &= ~(OBD_CONNECT_RMT_CLIENT |
- OBD_CONNECT_RMT_CLIENT_FORCE |
- OBD_CONNECT_MDS_CAPA |
- OBD_CONNECT_OSS_CAPA);
-}
-
-static int tgt_init_sec_level(struct ptlrpc_request *req)
+/** Assign high priority operations to the request if needed. */
+int tgt_hpreq_handler(struct ptlrpc_request *req)
{
- struct lu_target *tgt = class_exp2tgt(req->rq_export);
- char *client = libcfs_nid2str(req->rq_peer.nid);
- struct obd_connect_data *data, *reply;
- int rc = 0;
- bool remote;
+ struct tgt_session_info *tsi = tgt_ses_info(req->rq_svc_thread->t_env);
+ struct tgt_handler *h;
+ int rc;
ENTRY;
- data = req_capsule_client_get(&req->rq_pill, &RMF_CONNECT_DATA);
- reply = req_capsule_server_get(&req->rq_pill, &RMF_CONNECT_DATA);
- if (data == NULL || reply == NULL)
- RETURN(-EFAULT);
-
- /* connection from MDT is always trusted */
- if (req->rq_auth_usr_mdt) {
- tgt_init_sec_none(reply);
+ if (req->rq_export == NULL)
RETURN(0);
- }
- /* no GSS support case */
- if (!req->rq_auth_gss) {
- if (tgt->lut_sec_level > LUSTRE_SEC_NONE) {
- CWARN("client %s -> target %s does not use GSS, "
- "can not run under security level %d.\n",
- client, tgt_name(tgt), tgt->lut_sec_level);
- RETURN(-EACCES);
- } else {
- tgt_init_sec_none(reply);
- RETURN(0);
- }
- }
-
- /* old version case */
- if (unlikely(!(data->ocd_connect_flags & OBD_CONNECT_RMT_CLIENT) ||
- !(data->ocd_connect_flags & OBD_CONNECT_MDS_CAPA) ||
- !(data->ocd_connect_flags & OBD_CONNECT_OSS_CAPA))) {
- if (tgt->lut_sec_level > LUSTRE_SEC_NONE) {
- CWARN("client %s -> target %s uses old version, "
- "can not run under security level %d.\n",
- client, tgt_name(tgt), tgt->lut_sec_level);
- RETURN(-EACCES);
- } else {
- CWARN("client %s -> target %s uses old version, "
- "run under security level %d.\n",
- client, tgt_name(tgt), tgt->lut_sec_level);
- tgt_init_sec_none(reply);
- RETURN(0);
- }
- }
+ req_capsule_init(&req->rq_pill, req, RCL_SERVER);
+ tsi->tsi_pill = &req->rq_pill;
+ tsi->tsi_env = req->rq_svc_thread->t_env;
+ tsi->tsi_tgt = class_exp2tgt(req->rq_export);
+ tsi->tsi_exp = req->rq_export;
- remote = data->ocd_connect_flags & OBD_CONNECT_RMT_CLIENT_FORCE;
- if (remote) {
- if (!req->rq_auth_remote)
- CDEBUG(D_SEC, "client (local realm) %s -> target %s "
- "asked to be remote.\n", client, tgt_name(tgt));
- } else if (req->rq_auth_remote) {
- remote = true;
- CDEBUG(D_SEC, "client (remote realm) %s -> target %s is set "
- "as remote by default.\n", client, tgt_name(tgt));
- }
-
- if (remote) {
- if (!tgt->lut_oss_capa) {
- CDEBUG(D_SEC,
- "client %s -> target %s is set as remote,"
- " but OSS capabilities are not enabled: %d.\n",
- client, tgt_name(tgt), tgt->lut_oss_capa);
- RETURN(-EACCES);
- }
- } else {
- if (req->rq_auth_uid == INVALID_UID) {
- CDEBUG(D_SEC, "client %s -> target %s: user is not "
- "authenticated!\n", client, tgt_name(tgt));
- RETURN(-EACCES);
- }
+ h = tgt_handler_find_check(req);
+ if (IS_ERR(h)) {
+ rc = PTR_ERR(h);
+ RETURN(rc);
}
+ rc = tgt_request_preprocess(tsi, h, req);
+ if (unlikely(rc != 0))
+ RETURN(rc);
- switch (tgt->lut_sec_level) {
- case LUSTRE_SEC_NONE:
- if (remote) {
- CDEBUG(D_SEC,
- "client %s -> target %s is set as remote, "
- "can not run under security level %d.\n",
- client, tgt_name(tgt), tgt->lut_sec_level);
- RETURN(-EACCES);
- }
- tgt_init_sec_none(reply);
- break;
- case LUSTRE_SEC_REMOTE:
- if (!remote)
- tgt_init_sec_none(reply);
- break;
- case LUSTRE_SEC_ALL:
- if (remote)
- break;
- reply->ocd_connect_flags &= ~(OBD_CONNECT_RMT_CLIENT |
- OBD_CONNECT_RMT_CLIENT_FORCE);
- if (!tgt->lut_oss_capa)
- reply->ocd_connect_flags &= ~OBD_CONNECT_OSS_CAPA;
- if (!tgt->lut_mds_capa)
- reply->ocd_connect_flags &= ~OBD_CONNECT_MDS_CAPA;
- break;
- default:
- RETURN(-EINVAL);
- }
+ if (h->th_hp != NULL)
+ h->th_hp(tsi);
+ RETURN(0);
+}
+EXPORT_SYMBOL(tgt_hpreq_handler);
- RETURN(rc);
+void tgt_counter_incr(struct obd_export *exp, int opcode)
+{
+ lprocfs_counter_incr(exp->exp_obd->obd_stats, opcode);
+ if (exp->exp_nid_stats && exp->exp_nid_stats->nid_stats != NULL)
+ lprocfs_counter_incr(exp->exp_nid_stats->nid_stats, opcode);
}
+EXPORT_SYMBOL(tgt_counter_incr);
+
+/*
+ * Unified target generic handlers.
+ */
int tgt_connect_check_sptlrpc(struct ptlrpc_request *req, struct obd_export *exp)
{
spin_lock(&exp->exp_lock);
exp->exp_sp_peer = req->rq_sp_from;
exp->exp_flvr = flvr;
+
+ /* when on mgs, if no restriction is set, or if client
+ * is loopback, allow any flavor */
+ if ((strcmp(exp->exp_obd->obd_type->typ_name,
+ LUSTRE_MGS_NAME) == 0) &&
+ (exp->exp_flvr.sf_rpc == SPTLRPC_FLVR_NULL ||
+ LNET_NETTYP(LNET_NIDNET(exp->exp_connection->c_peer.nid))
+ == LOLND))
+ exp->exp_flvr.sf_rpc = SPTLRPC_FLVR_ANY;
+
if (exp->exp_flvr.sf_rpc != SPTLRPC_FLVR_ANY &&
exp->exp_flvr.sf_rpc != req->rq_flvr.sf_rpc) {
CERROR("%s: unauthorized rpc flavor %x from %s, "
return rc;
}
-int tgt_adapt_sptlrpc_conf(struct lu_target *tgt, int initial)
+int tgt_adapt_sptlrpc_conf(struct lu_target *tgt)
{
struct sptlrpc_rule_set tmp_rset;
int rc;
+ if (unlikely(tgt == NULL)) {
+ CERROR("No target passed");
+ return -EINVAL;
+ }
+
sptlrpc_rule_set_init(&tmp_rset);
- rc = sptlrpc_conf_target_get_rules(tgt->lut_obd, &tmp_rset, initial);
+ rc = sptlrpc_conf_target_get_rules(tgt->lut_obd, &tmp_rset);
if (rc) {
CERROR("%s: failed get sptlrpc rules: rc = %d\n",
tgt_name(tgt), rc);
ENTRY;
- rc = tgt_init_sec_level(req);
- if (rc != 0)
- GOTO(out, rc);
-
/* XXX: better to call this check right after getting new export but
* before last_rcvd slot allocation to avoid server load upon insecure
* connects. This is to be fixed after unifiyng all targets.
reply = req_capsule_server_get(tsi->tsi_pill, &RMF_CONNECT_DATA);
spin_lock(&tsi->tsi_exp->exp_lock);
*exp_connect_flags_ptr(tsi->tsi_exp) = reply->ocd_connect_flags;
+ if (reply->ocd_connect_flags & OBD_CONNECT_FLAGS2)
+ *exp_connect_flags2_ptr(tsi->tsi_exp) =
+ reply->ocd_connect_flags2;
tsi->tsi_exp->exp_connect_data.ocd_brw_size = reply->ocd_brw_size;
spin_unlock(&tsi->tsi_exp->exp_lock);
{
return err_serious(-EOPNOTSUPP);
}
-EXPORT_SYMBOL(tgt_obd_log_cancel);
-int tgt_obd_qc_callback(struct tgt_session_info *tsi)
+int tgt_send_buffer(struct tgt_session_info *tsi, struct lu_rdbuf *rdbuf)
{
- return err_serious(-EOPNOTSUPP);
+ struct tgt_thread_info *tti = tgt_th_info(tsi->tsi_env);
+ struct ptlrpc_request *req = tgt_ses_req(tsi);
+ struct obd_export *exp = req->rq_export;
+ struct ptlrpc_bulk_desc *desc;
+ struct l_wait_info *lwi = &tti->tti_u.update.tti_wait_info;
+ int i;
+ int rc;
+
+ ENTRY;
+
+ desc = ptlrpc_prep_bulk_exp(req, rdbuf->rb_nbufs, 1,
+ PTLRPC_BULK_PUT_SOURCE | PTLRPC_BULK_BUF_KVEC,
+ MDS_BULK_PORTAL, &ptlrpc_bulk_kvec_ops);
+ if (desc == NULL)
+ RETURN(-ENOMEM);
+
+ for (i = 0; i < rdbuf->rb_nbufs; i++)
+ desc->bd_frag_ops->add_iov_frag(desc,
+ rdbuf->rb_bufs[i].lb_buf,
+ rdbuf->rb_bufs[i].lb_len);
+
+ rc = target_bulk_io(exp, desc, lwi);
+ ptlrpc_free_bulk(desc);
+ RETURN(rc);
}
-EXPORT_SYMBOL(tgt_obd_qc_callback);
+EXPORT_SYMBOL(tgt_send_buffer);
int tgt_sendpage(struct tgt_session_info *tsi, struct lu_rdpg *rdpg, int nob)
{
ENTRY;
- desc = ptlrpc_prep_bulk_exp(req, rdpg->rp_npages, 1, BULK_PUT_SOURCE,
- MDS_BULK_PORTAL);
+ desc = ptlrpc_prep_bulk_exp(req, rdpg->rp_npages, 1,
+ PTLRPC_BULK_PUT_SOURCE |
+ PTLRPC_BULK_BUF_KIOV,
+ MDS_BULK_PORTAL,
+ &ptlrpc_bulk_kiov_pin_ops);
if (desc == NULL)
RETURN(-ENOMEM);
if (!(exp_connect_flags(exp) & OBD_CONNECT_BRW_SIZE))
- /* old client requires reply size in it's PAGE_CACHE_SIZE,
+ /* old client requires reply size in it's PAGE_SIZE,
* which is rdpg->rp_count */
nob = rdpg->rp_count;
for (i = 0, tmpcount = nob; i < rdpg->rp_npages && tmpcount > 0;
i++, tmpcount -= tmpsize) {
- tmpsize = min_t(int, tmpcount, PAGE_CACHE_SIZE);
- ptlrpc_prep_bulk_page_pin(desc, rdpg->rp_pages[i], 0, tmpsize);
+ tmpsize = min_t(int, tmpcount, PAGE_SIZE);
+ desc->bd_frag_ops->add_kiov_frag(desc, rdpg->rp_pages[i], 0,
+ tmpsize);
}
LASSERT(desc->bd_nob == nob);
rc = target_bulk_io(exp, desc, lwi);
- ptlrpc_free_bulk_pin(desc);
+ ptlrpc_free_bulk(desc);
RETURN(rc);
}
EXPORT_SYMBOL(tgt_sendpage);
/*
* OBD_IDX_READ handler
*/
-int tgt_obd_idx_read(struct tgt_session_info *tsi)
+static int tgt_obd_idx_read(struct tgt_session_info *tsi)
{
struct tgt_thread_info *tti = tgt_th_info(tsi->tsi_env);
struct lu_rdpg *rdpg = &tti->tti_u.rdpg.tti_rdpg;
GOTO(out, rc = -EFAULT);
rdpg->rp_count = min_t(unsigned int, req_ii->ii_count << LU_PAGE_SHIFT,
exp_max_brw_size(tsi->tsi_exp));
- rdpg->rp_npages = (rdpg->rp_count + PAGE_CACHE_SIZE -1) >> PAGE_CACHE_SHIFT;
+ rdpg->rp_npages = (rdpg->rp_count + PAGE_SIZE - 1) >> PAGE_SHIFT;
/* allocate pages to store the containers */
OBD_ALLOC(rdpg->rp_pages, rdpg->rp_npages * sizeof(rdpg->rp_pages[0]));
if (rdpg->rp_pages == NULL)
GOTO(out, rc = -ENOMEM);
for (i = 0; i < rdpg->rp_npages; i++) {
- rdpg->rp_pages[i] = alloc_page(GFP_IOFS);
+ rdpg->rp_pages[i] = alloc_page(GFP_NOFS);
if (rdpg->rp_pages[i] == NULL)
GOTO(out, rc = -ENOMEM);
}
}
return rc;
}
-EXPORT_SYMBOL(tgt_obd_idx_read);
struct tgt_handler tgt_obd_handlers[] = {
TGT_OBD_HDL (0, OBD_PING, tgt_obd_ping),
-TGT_OBD_HDL_VAR(0, OBD_LOG_CANCEL, tgt_obd_log_cancel),
-TGT_OBD_HDL_VAR(0, OBD_QC_CALLBACK, tgt_obd_qc_callback),
TGT_OBD_HDL (0, OBD_IDX_READ, tgt_obd_idx_read)
};
EXPORT_SYMBOL(tgt_obd_handlers);
int tgt_sync(const struct lu_env *env, struct lu_target *tgt,
- struct dt_object *obj)
+ struct dt_object *obj, __u64 start, __u64 end)
{
int rc = 0;
rc = dt_sync(env, tgt->lut_bottom);
} else if (dt_version_get(env, obj) >
tgt->lut_obd->obd_last_committed) {
- rc = dt_object_sync(env, obj);
+ rc = dt_object_sync(env, obj, start, end);
}
+ atomic_inc(&tgt->lut_sync_count);
RETURN(rc);
}
* Unified target DLM handlers.
*/
-/* Ensure that data and metadata are synced to the disk when lock is cancelled
- * (if requested) */
-int tgt_blocking_ast(struct ldlm_lock *lock, struct ldlm_lock_desc *desc,
- void *data, int flag)
+/**
+ * Unified target BAST
+ *
+ * Ensure data and metadata are synced to disk when lock is canceled if Sync on
+ * Cancel (SOC) is enabled. If it's extent lock, normally sync obj is enough,
+ * but if it's cross-MDT lock, because remote object version is not set, a
+ * filesystem sync is needed.
+ *
+ * \param lock server side lock
+ * \param desc lock desc
+ * \param data ldlm_cb_set_arg
+ * \param flag indicates whether this cancelling or blocking callback
+ * \retval 0 on success
+ * \retval negative number on error
+ */
+static int tgt_blocking_ast(struct ldlm_lock *lock, struct ldlm_lock_desc *desc,
+ void *data, int flag)
{
struct lu_env env;
struct lu_target *tgt;
- struct dt_object *obj;
+ struct dt_object *obj = NULL;
struct lu_fid fid;
int rc = 0;
tgt = class_exp2tgt(lock->l_export);
+ if (unlikely(tgt == NULL)) {
+ CDEBUG(D_ERROR, "%s: No target for connected export\n",
+ class_exp2obd(lock->l_export)->obd_name);
+ RETURN(-EINVAL);
+ }
+
if (flag == LDLM_CB_CANCELING &&
- (lock->l_granted_mode & (LCK_PW | LCK_GROUP)) &&
+ (lock->l_granted_mode & (LCK_EX | LCK_PW | LCK_GROUP)) &&
(tgt->lut_sync_lock_cancel == ALWAYS_SYNC_ON_CANCEL ||
(tgt->lut_sync_lock_cancel == BLOCKING_SYNC_ON_CANCEL &&
- lock->l_flags & LDLM_FL_CBPENDING))) {
+ ldlm_is_cbpending(lock))) &&
+ ((exp_connect_flags(lock->l_export) & OBD_CONNECT_MDS_MDS) ||
+ lock->l_resource->lr_type == LDLM_EXTENT)) {
+ __u64 start = 0;
+ __u64 end = OBD_OBJECT_EOF;
+
rc = lu_env_init(&env, LCT_DT_THREAD);
if (unlikely(rc != 0))
- RETURN(rc);
+ GOTO(err, rc);
- ost_fid_from_resid(&fid, &lock->l_resource->lr_name);
- obj = dt_locate(&env, tgt->lut_bottom, &fid);
- if (IS_ERR(obj))
- GOTO(err_env, rc = PTR_ERR(obj));
+ ost_fid_from_resid(&fid, &lock->l_resource->lr_name,
+ tgt->lut_lsd.lsd_osd_index);
- if (!dt_object_exists(obj))
- GOTO(err_put, rc = -ENOENT);
+ if (lock->l_resource->lr_type == LDLM_EXTENT) {
+ obj = dt_locate(&env, tgt->lut_bottom, &fid);
+ if (IS_ERR(obj))
+ GOTO(err_env, rc = PTR_ERR(obj));
- rc = tgt_sync(&env, tgt, obj);
+ if (!dt_object_exists(obj))
+ GOTO(err_put, rc = -ENOENT);
+
+ start = lock->l_policy_data.l_extent.start;
+ end = lock->l_policy_data.l_extent.end;
+ }
+
+ rc = tgt_sync(&env, tgt, obj, start, end);
if (rc < 0) {
- CERROR("%s: sync failed on lock cancel: rc = %d\n",
- tgt_name(tgt), rc);
+ CERROR("%s: syncing "DFID" (%llu-%llu) on lock "
+ "cancel: rc = %d\n",
+ tgt_name(tgt), PFID(&fid),
+ lock->l_policy_data.l_extent.start,
+ lock->l_policy_data.l_extent.end, rc);
}
err_put:
- lu_object_put(&env, &obj->do_lu);
+ if (obj != NULL)
+ dt_object_put(&env, obj);
err_env:
lu_env_fini(&env);
}
-
+err:
rc = ldlm_server_blocking_ast(lock, desc, data, flag);
RETURN(rc);
}
-struct ldlm_callback_suite tgt_dlm_cbs = {
+static struct ldlm_callback_suite tgt_dlm_cbs = {
.lcs_completion = ldlm_server_completion_ast,
.lcs_blocking = tgt_blocking_ast,
.lcs_glimpse = ldlm_server_glimpse_ast
if (rc)
RETURN(err_serious(rc));
+ switch (LUT_FAIL_CLASS(tsi->tsi_reply_fail_id)) {
+ case LUT_FAIL_MDT:
+ tsi->tsi_reply_fail_id = OBD_FAIL_MDS_LDLM_REPLY_NET;
+ break;
+ case LUT_FAIL_OST:
+ tsi->tsi_reply_fail_id = OBD_FAIL_OST_LDLM_REPLY_NET;
+ break;
+ case LUT_FAIL_MGT:
+ tsi->tsi_reply_fail_id = OBD_FAIL_MGS_LDLM_REPLY_NET;
+ break;
+ default:
+ tsi->tsi_reply_fail_id = OBD_FAIL_LDLM_REPLY;
+ break;
+ }
RETURN(req->rq_status);
}
EXPORT_SYMBOL(tgt_enqueue);
RETURN(req->rq_status);
}
-EXPORT_SYMBOL(tgt_convert);
int tgt_bl_callback(struct tgt_session_info *tsi)
{
return err_serious(-EOPNOTSUPP);
}
-EXPORT_SYMBOL(tgt_bl_callback);
int tgt_cp_callback(struct tgt_session_info *tsi)
{
return err_serious(-EOPNOTSUPP);
}
-EXPORT_SYMBOL(tgt_cp_callback);
/* generic LDLM target handler */
struct tgt_handler tgt_dlm_handlers[] = {
TGT_DLM_HDL (HABEO_CLAVIS, LDLM_ENQUEUE, tgt_enqueue),
-TGT_DLM_HDL_VAR(HABEO_CLAVIS, LDLM_CONVERT, tgt_convert),
+TGT_DLM_HDL (HABEO_CLAVIS, LDLM_CONVERT, tgt_convert),
TGT_DLM_HDL_VAR(0, LDLM_BL_CALLBACK, tgt_bl_callback),
TGT_DLM_HDL_VAR(0, LDLM_CP_CALLBACK, tgt_cp_callback)
};
}
EXPORT_SYMBOL(tgt_llog_open);
-int tgt_llog_close(struct tgt_session_info *tsi)
-{
- int rc;
-
- ENTRY;
-
- rc = llog_origin_handle_close(tgt_ses_req(tsi));
-
- RETURN(rc);
-}
-EXPORT_SYMBOL(tgt_llog_close);
-
-
-int tgt_llog_destroy(struct tgt_session_info *tsi)
-{
- int rc;
-
- ENTRY;
-
- rc = llog_origin_handle_destroy(tgt_ses_req(tsi));
-
- RETURN(rc);
-}
-EXPORT_SYMBOL(tgt_llog_destroy);
-
int tgt_llog_read_header(struct tgt_session_info *tsi)
{
int rc;
TGT_LLOG_HDL (0, LLOG_ORIGIN_HANDLE_NEXT_BLOCK, tgt_llog_next_block),
TGT_LLOG_HDL (0, LLOG_ORIGIN_HANDLE_READ_HEADER, tgt_llog_read_header),
TGT_LLOG_HDL (0, LLOG_ORIGIN_HANDLE_PREV_BLOCK, tgt_llog_prev_block),
-TGT_LLOG_HDL (0, LLOG_ORIGIN_HANDLE_DESTROY, tgt_llog_destroy),
-TGT_LLOG_HDL_VAR(0, LLOG_ORIGIN_HANDLE_CLOSE, tgt_llog_close),
};
EXPORT_SYMBOL(tgt_llog_handlers);
* sec context handlers
*/
/* XXX: Implement based on mdt_sec_ctx_handle()? */
-int tgt_sec_ctx_handle(struct tgt_session_info *tsi)
+static int tgt_sec_ctx_handle(struct tgt_session_info *tsi)
{
return 0;
}
};
EXPORT_SYMBOL(tgt_sec_ctx_handlers);
+int (*tgt_lfsck_in_notify_local)(const struct lu_env *env,
+ struct dt_device *key,
+ struct lfsck_req_local *lrl,
+ struct thandle *th) = NULL;
+
+void tgt_register_lfsck_in_notify_local(int (*notify)(const struct lu_env *,
+ struct dt_device *,
+ struct lfsck_req_local *,
+ struct thandle *))
+{
+ tgt_lfsck_in_notify_local = notify;
+}
+EXPORT_SYMBOL(tgt_register_lfsck_in_notify_local);
+
+int (*tgt_lfsck_in_notify)(const struct lu_env *env,
+ struct dt_device *key,
+ struct lfsck_request *lr) = NULL;
+
+void tgt_register_lfsck_in_notify(int (*notify)(const struct lu_env *,
+ struct dt_device *,
+ struct lfsck_request *))
+{
+ tgt_lfsck_in_notify = notify;
+}
+EXPORT_SYMBOL(tgt_register_lfsck_in_notify);
+
+static int (*tgt_lfsck_query)(const struct lu_env *env,
+ struct dt_device *key,
+ struct lfsck_request *req,
+ struct lfsck_reply *rep,
+ struct lfsck_query *que) = NULL;
+
+void tgt_register_lfsck_query(int (*query)(const struct lu_env *,
+ struct dt_device *,
+ struct lfsck_request *,
+ struct lfsck_reply *,
+ struct lfsck_query *))
+{
+ tgt_lfsck_query = query;
+}
+EXPORT_SYMBOL(tgt_register_lfsck_query);
+
+/* LFSCK request handlers */
+static int tgt_handle_lfsck_notify(struct tgt_session_info *tsi)
+{
+ const struct lu_env *env = tsi->tsi_env;
+ struct dt_device *key = tsi->tsi_tgt->lut_bottom;
+ struct lfsck_request *lr;
+ int rc;
+ ENTRY;
+
+ lr = req_capsule_client_get(tsi->tsi_pill, &RMF_LFSCK_REQUEST);
+ if (lr == NULL)
+ RETURN(-EPROTO);
+
+ rc = tgt_lfsck_in_notify(env, key, lr);
+
+ RETURN(rc);
+}
+
+static int tgt_handle_lfsck_query(struct tgt_session_info *tsi)
+{
+ struct lfsck_request *request;
+ struct lfsck_reply *reply;
+ int rc;
+ ENTRY;
+
+ request = req_capsule_client_get(tsi->tsi_pill, &RMF_LFSCK_REQUEST);
+ if (request == NULL)
+ RETURN(-EPROTO);
+
+ reply = req_capsule_server_get(tsi->tsi_pill, &RMF_LFSCK_REPLY);
+ if (reply == NULL)
+ RETURN(-ENOMEM);
+
+ rc = tgt_lfsck_query(tsi->tsi_env, tsi->tsi_tgt->lut_bottom,
+ request, reply, NULL);
+
+ RETURN(rc < 0 ? rc : 0);
+}
+
+struct tgt_handler tgt_lfsck_handlers[] = {
+TGT_LFSCK_HDL(HABEO_REFERO, LFSCK_NOTIFY, tgt_handle_lfsck_notify),
+TGT_LFSCK_HDL(HABEO_REFERO, LFSCK_QUERY, tgt_handle_lfsck_query),
+};
+EXPORT_SYMBOL(tgt_lfsck_handlers);
+
/*
* initialize per-thread page pool (bug 5137).
*/
EXIT;
}
EXPORT_SYMBOL(tgt_io_thread_done);
+
+/**
+ * Helper function for getting Data-on-MDT file server DLM lock
+ * if asked by client.
+ */
+int tgt_mdt_data_lock(struct ldlm_namespace *ns, struct ldlm_res_id *res_id,
+ struct lustre_handle *lh, int mode, __u64 *flags)
+{
+ union ldlm_policy_data policy = {
+ .l_inodebits.bits = MDS_INODELOCK_DOM,
+ };
+ int rc;
+
+ ENTRY;
+
+ LASSERT(lh != NULL);
+ LASSERT(ns != NULL);
+ LASSERT(!lustre_handle_is_used(lh));
+
+ rc = ldlm_cli_enqueue_local(NULL, ns, res_id, LDLM_IBITS, &policy, mode,
+ flags, ldlm_blocking_ast,
+ ldlm_completion_ast, ldlm_glimpse_ast,
+ NULL, 0, LVB_T_NONE, NULL, lh);
+
+ RETURN(rc == ELDLM_OK ? 0 : -EIO);
+}
+EXPORT_SYMBOL(tgt_mdt_data_lock);
+
+void tgt_mdt_data_unlock(struct lustre_handle *lh, enum ldlm_mode mode)
+{
+ LASSERT(lustre_handle_is_used(lh));
+ ldlm_lock_decref(lh, mode);
+}
+EXPORT_SYMBOL(tgt_mdt_data_unlock);
+
/**
* Helper function for getting server side [start, start+count] DLM lock
* if asked by client.
*/
-int tgt_extent_lock(struct ldlm_namespace *ns, struct ldlm_res_id *res_id,
- __u64 start, __u64 end, struct lustre_handle *lh,
- int mode, __u64 *flags)
+int tgt_extent_lock(const struct lu_env *env, struct ldlm_namespace *ns,
+ struct ldlm_res_id *res_id, __u64 start, __u64 end,
+ struct lustre_handle *lh, int mode, __u64 *flags)
{
- ldlm_policy_data_t policy;
- int rc;
+ union ldlm_policy_data policy;
+ int rc;
ENTRY;
LASSERT(!lustre_handle_is_used(lh));
policy.l_extent.gid = 0;
- policy.l_extent.start = start & CFS_PAGE_MASK;
+ policy.l_extent.start = start & PAGE_MASK;
/*
* If ->o_blocks is EOF it means "lock till the end of the file".
if (end == OBD_OBJECT_EOF || end < start)
policy.l_extent.end = OBD_OBJECT_EOF;
else
- policy.l_extent.end = end | ~CFS_PAGE_MASK;
+ policy.l_extent.end = end | ~PAGE_MASK;
- rc = ldlm_cli_enqueue_local(ns, res_id, LDLM_EXTENT, &policy, mode,
- flags, ldlm_blocking_ast,
+ rc = ldlm_cli_enqueue_local(env, ns, res_id, LDLM_EXTENT, &policy,
+ mode, flags, ldlm_blocking_ast,
ldlm_completion_ast, ldlm_glimpse_ast,
NULL, 0, LVB_T_NONE, NULL, lh);
RETURN(rc == ELDLM_OK ? 0 : -EIO);
}
EXPORT_SYMBOL(tgt_extent_lock);
-void tgt_extent_unlock(struct lustre_handle *lh, ldlm_mode_t mode)
+void tgt_extent_unlock(struct lustre_handle *lh, enum ldlm_mode mode)
{
LASSERT(lustre_handle_is_used(lh));
ldlm_lock_decref(lh, mode);
}
EXPORT_SYMBOL(tgt_extent_unlock);
-int tgt_brw_lock(struct ldlm_namespace *ns, struct ldlm_res_id *res_id,
- struct obd_ioobj *obj, struct niobuf_remote *nb,
- struct lustre_handle *lh, int mode)
+static int tgt_brw_lock(const struct lu_env *env, struct obd_export *exp,
+ struct ldlm_res_id *res_id, struct obd_ioobj *obj,
+ struct niobuf_remote *nb, struct lustre_handle *lh,
+ enum ldlm_mode mode)
{
+ struct ldlm_namespace *ns = exp->exp_obd->obd_namespace;
__u64 flags = 0;
int nrbufs = obj->ioo_bufcnt;
int i;
+ int rc;
ENTRY;
LASSERT(mode == LCK_PR || mode == LCK_PW);
LASSERT(!lustre_handle_is_used(lh));
- if (nrbufs == 0 || !(nb[0].flags & OBD_BRW_SRVLOCK))
+ if (ns->ns_obd->obd_recovering)
+ RETURN(0);
+
+ if (nrbufs == 0 || !(nb[0].rnb_flags & OBD_BRW_SRVLOCK))
RETURN(0);
for (i = 1; i < nrbufs; i++)
- if (!(nb[i].flags & OBD_BRW_SRVLOCK))
+ if (!(nb[i].rnb_flags & OBD_BRW_SRVLOCK))
RETURN(-EFAULT);
- RETURN(tgt_extent_lock(ns, res_id, nb[0].offset,
- nb[nrbufs - 1].offset + nb[nrbufs - 1].len - 1,
- lh, mode, &flags));
+ /* MDT IO for data-on-mdt */
+ if (exp->exp_connect_data.ocd_connect_flags & OBD_CONNECT_IBITS)
+ rc = tgt_mdt_data_lock(ns, res_id, lh, mode, &flags);
+ else
+ rc = tgt_extent_lock(env, ns, res_id, nb[0].rnb_offset,
+ nb[nrbufs - 1].rnb_offset +
+ nb[nrbufs - 1].rnb_len - 1,
+ lh, mode, &flags);
+ RETURN(rc);
}
-EXPORT_SYMBOL(tgt_brw_lock);
-void tgt_brw_unlock(struct obd_ioobj *obj, struct niobuf_remote *niob,
- struct lustre_handle *lh, int mode)
+static void tgt_brw_unlock(struct obd_ioobj *obj, struct niobuf_remote *niob,
+ struct lustre_handle *lh, enum ldlm_mode mode)
{
ENTRY;
LASSERT(mode == LCK_PR || mode == LCK_PW);
- LASSERT((obj->ioo_bufcnt > 0 && (niob[0].flags & OBD_BRW_SRVLOCK)) ==
+ LASSERT((obj->ioo_bufcnt > 0 &&
+ (niob[0].rnb_flags & OBD_BRW_SRVLOCK)) ==
lustre_handle_is_used(lh));
+
if (lustre_handle_is_used(lh))
tgt_extent_unlock(lh, mode);
EXIT;
}
-EXPORT_SYMBOL(tgt_brw_unlock);
-
-static __u32 tgt_checksum_bulk(struct lu_target *tgt,
- struct ptlrpc_bulk_desc *desc, int opc,
- cksum_type_t cksum_type)
+static int tgt_checksum_niobuf(struct lu_target *tgt,
+ struct niobuf_local *local_nb, int npages,
+ int opc, enum cksum_types cksum_type,
+ __u32 *cksum)
{
- struct cfs_crypto_hash_desc *hdesc;
+ struct ahash_request *req;
unsigned int bufsize;
int i, err;
unsigned char cfs_alg = cksum_obd2cfs(cksum_type);
- __u32 cksum;
- hdesc = cfs_crypto_hash_init(cfs_alg, NULL, 0);
- if (IS_ERR(hdesc)) {
+ req = cfs_crypto_hash_init(cfs_alg, NULL, 0);
+ if (IS_ERR(req)) {
CERROR("%s: unable to initialize checksum hash %s\n",
tgt_name(tgt), cfs_crypto_hash_name(cfs_alg));
- return PTR_ERR(hdesc);
+ return PTR_ERR(req);
}
CDEBUG(D_INFO, "Checksum for algo %s\n", cfs_crypto_hash_name(cfs_alg));
- for (i = 0; i < desc->bd_iov_count; i++) {
+ for (i = 0; i < npages; i++) {
/* corrupt the data before we compute the checksum, to
* simulate a client->OST data error */
if (i == 0 && opc == OST_WRITE &&
OBD_FAIL_CHECK(OBD_FAIL_OST_CHECKSUM_RECEIVE)) {
- int off = desc->bd_iov[i].kiov_offset & ~CFS_PAGE_MASK;
- int len = desc->bd_iov[i].kiov_len;
+ int off = local_nb[i].lnb_page_offset & ~PAGE_MASK;
+ int len = local_nb[i].lnb_len;
struct page *np = tgt_page_to_corrupt;
- char *ptr = kmap(desc->bd_iov[i].kiov_page) + off;
if (np) {
- char *ptr2 = kmap(np) + off;
+ char *ptr = ll_kmap_atomic(local_nb[i].lnb_page,
+ KM_USER0);
+ char *ptr2 = page_address(np);
+
+ memcpy(ptr2 + off, ptr + off, len);
+ memcpy(ptr2 + off, "bad3", min(4, len));
+ ll_kunmap_atomic(ptr, KM_USER0);
+
+ /* LU-8376 to preserve original index for
+ * display in dump_all_bulk_pages() */
+ np->index = i;
- memcpy(ptr2, ptr, len);
- memcpy(ptr2, "bad3", min(4, len));
- kunmap(np);
- desc->bd_iov[i].kiov_page = np;
+ cfs_crypto_hash_update_page(req, np, off,
+ len);
+ continue;
} else {
CERROR("%s: can't alloc page for corruption\n",
tgt_name(tgt));
}
}
- cfs_crypto_hash_update_page(hdesc, desc->bd_iov[i].kiov_page,
- desc->bd_iov[i].kiov_offset & ~CFS_PAGE_MASK,
- desc->bd_iov[i].kiov_len);
+ cfs_crypto_hash_update_page(req, local_nb[i].lnb_page,
+ local_nb[i].lnb_page_offset & ~PAGE_MASK,
+ local_nb[i].lnb_len);
/* corrupt the data after we compute the checksum, to
* simulate an OST->client data error */
if (i == 0 && opc == OST_READ &&
OBD_FAIL_CHECK(OBD_FAIL_OST_CHECKSUM_SEND)) {
- int off = desc->bd_iov[i].kiov_offset & ~CFS_PAGE_MASK;
- int len = desc->bd_iov[i].kiov_len;
+ int off = local_nb[i].lnb_page_offset & ~PAGE_MASK;
+ int len = local_nb[i].lnb_len;
+ struct page *np = tgt_page_to_corrupt;
+
+ if (np) {
+ char *ptr = ll_kmap_atomic(local_nb[i].lnb_page,
+ KM_USER0);
+ char *ptr2 = page_address(np);
+
+ memcpy(ptr2 + off, ptr + off, len);
+ memcpy(ptr2 + off, "bad4", min(4, len));
+ ll_kunmap_atomic(ptr, KM_USER0);
+
+ /* LU-8376 to preserve original index for
+ * display in dump_all_bulk_pages() */
+ np->index = i;
+
+ cfs_crypto_hash_update_page(req, np, off,
+ len);
+ continue;
+ } else {
+ CERROR("%s: can't alloc page for corruption\n",
+ tgt_name(tgt));
+ }
+ }
+ }
+
+ bufsize = sizeof(*cksum);
+ err = cfs_crypto_hash_final(req, (unsigned char *)cksum, &bufsize);
+
+ return 0;
+}
+
+char dbgcksum_file_name[PATH_MAX];
+
+static void dump_all_bulk_pages(struct obdo *oa, int count,
+ struct niobuf_local *local_nb,
+ __u32 server_cksum, __u32 client_cksum)
+{
+ struct file *filp;
+ int rc, i;
+ unsigned int len;
+ char *buf;
+
+ /* will only keep dump of pages on first error for the same range in
+ * file/fid, not during the resends/retries. */
+ snprintf(dbgcksum_file_name, sizeof(dbgcksum_file_name),
+ "%s-checksum_dump-ost-"DFID":[%llu-%llu]-%x-%x",
+ (strncmp(libcfs_debug_file_path_arr, "NONE", 4) != 0 ?
+ libcfs_debug_file_path_arr :
+ LIBCFS_DEBUG_FILE_PATH_DEFAULT),
+ oa->o_valid & OBD_MD_FLFID ? oa->o_parent_seq : (__u64)0,
+ oa->o_valid & OBD_MD_FLFID ? oa->o_parent_oid : 0,
+ oa->o_valid & OBD_MD_FLFID ? oa->o_parent_ver : 0,
+ local_nb[0].lnb_file_offset,
+ local_nb[count-1].lnb_file_offset +
+ local_nb[count-1].lnb_len - 1, client_cksum, server_cksum);
+ filp = filp_open(dbgcksum_file_name,
+ O_CREAT | O_EXCL | O_WRONLY | O_LARGEFILE, 0600);
+ if (IS_ERR(filp)) {
+ rc = PTR_ERR(filp);
+ if (rc == -EEXIST)
+ CDEBUG(D_INFO, "%s: can't open to dump pages with "
+ "checksum error: rc = %d\n", dbgcksum_file_name,
+ rc);
+ else
+ CERROR("%s: can't open to dump pages with checksum "
+ "error: rc = %d\n", dbgcksum_file_name, rc);
+ return;
+ }
+
+ for (i = 0; i < count; i++) {
+ len = local_nb[i].lnb_len;
+ buf = kmap(local_nb[i].lnb_page);
+ while (len != 0) {
+ rc = cfs_kernel_write(filp, buf, len, &filp->f_pos);
+ if (rc < 0) {
+ CERROR("%s: wanted to write %u but got %d "
+ "error\n", dbgcksum_file_name, len, rc);
+ break;
+ }
+ len -= rc;
+ buf += rc;
+ CDEBUG(D_INFO, "%s: wrote %d bytes\n",
+ dbgcksum_file_name, rc);
+ }
+ kunmap(local_nb[i].lnb_page);
+ }
+
+ rc = ll_vfs_fsync_range(filp, 0, LLONG_MAX, 1);
+ if (rc)
+ CERROR("%s: sync returns %d\n", dbgcksum_file_name, rc);
+ filp_close(filp, NULL);
+ return;
+}
+
+static int check_read_checksum(struct niobuf_local *local_nb, int npages,
+ struct obd_export *exp, struct obdo *oa,
+ const struct lnet_process_id *peer,
+ __u32 client_cksum, __u32 server_cksum,
+ enum cksum_types server_cksum_type)
+{
+ char *msg;
+ enum cksum_types cksum_type;
+ loff_t start, end;
+
+ /* unlikely to happen and only if resend does not occur due to cksum
+ * control failure on Client */
+ if (unlikely(server_cksum == client_cksum)) {
+ CDEBUG(D_PAGE, "checksum %x confirmed upon retry\n",
+ client_cksum);
+ return 0;
+ }
+
+ if (exp->exp_obd->obd_checksum_dump)
+ dump_all_bulk_pages(oa, npages, local_nb, server_cksum,
+ client_cksum);
+
+ cksum_type = obd_cksum_type_unpack(oa->o_valid & OBD_MD_FLFLAGS ?
+ oa->o_flags : 0);
+
+ if (cksum_type != server_cksum_type)
+ msg = "the server may have not used the checksum type specified"
+ " in the original request - likely a protocol problem";
+ else
+ msg = "should have changed on the client or in transit";
+
+ start = local_nb[0].lnb_file_offset;
+ end = local_nb[npages-1].lnb_file_offset +
+ local_nb[npages-1].lnb_len - 1;
+
+ LCONSOLE_ERROR_MSG(0x132, "%s: BAD READ CHECKSUM: %s: from %s inode "
+ DFID " object "DOSTID" extent [%llu-%llu], client returned csum"
+ " %x (type %x), server csum %x (type %x)\n",
+ exp->exp_obd->obd_name,
+ msg, libcfs_nid2str(peer->nid),
+ oa->o_valid & OBD_MD_FLFID ? oa->o_parent_seq : 0ULL,
+ oa->o_valid & OBD_MD_FLFID ? oa->o_parent_oid : 0,
+ oa->o_valid & OBD_MD_FLFID ? oa->o_parent_ver : 0,
+ POSTID(&oa->o_oi),
+ start, end, client_cksum, cksum_type, server_cksum,
+ server_cksum_type);
+
+ return 1;
+}
+
+static int tgt_pages2shortio(struct niobuf_local *local, int npages,
+ unsigned char *buf, int size)
+{
+ int i, off, len, copied = size;
+ char *ptr;
+
+ for (i = 0; i < npages; i++) {
+ off = local[i].lnb_page_offset & ~PAGE_MASK;
+ len = local[i].lnb_len;
+
+ CDEBUG(D_PAGE, "index %d offset = %d len = %d left = %d\n",
+ i, off, len, size);
+ if (len > size)
+ return -EINVAL;
+
+ ptr = ll_kmap_atomic(local[i].lnb_page, KM_USER0);
+ memcpy(buf + off, ptr, len);
+ ll_kunmap_atomic(ptr, KM_USER0);
+ buf += len;
+ size -= len;
+ }
+ return copied - size;
+}
+
+static int tgt_checksum_niobuf_t10pi(struct lu_target *tgt,
+ struct niobuf_local *local_nb,
+ int npages, int opc,
+ obd_dif_csum_fn *fn,
+ int sector_size,
+ u32 *check_sum)
+{
+ enum cksum_types t10_cksum_type = tgt->lut_dt_conf.ddp_t10_cksum_type;
+ unsigned char cfs_alg = cksum_obd2cfs(OBD_CKSUM_T10_TOP);
+ const char *obd_name = tgt->lut_obd->obd_name;
+ struct ahash_request *req;
+ unsigned int bufsize;
+ unsigned char *buffer;
+ struct page *__page;
+ __u16 *guard_start;
+ int guard_number;
+ int used_number = 0;
+ __u32 cksum;
+ int rc = 0;
+ int used;
+ int i;
+
+ __page = alloc_page(GFP_KERNEL);
+ if (__page == NULL)
+ return -ENOMEM;
+
+ req = cfs_crypto_hash_init(cfs_alg, NULL, 0);
+ if (IS_ERR(req)) {
+ CERROR("%s: unable to initialize checksum hash %s\n",
+ tgt_name(tgt), cfs_crypto_hash_name(cfs_alg));
+ return PTR_ERR(req);
+ }
+
+ buffer = kmap(__page);
+ guard_start = (__u16 *)buffer;
+ guard_number = PAGE_SIZE / sizeof(*guard_start);
+ for (i = 0; i < npages; i++) {
+ /* corrupt the data before we compute the checksum, to
+ * simulate a client->OST data error */
+ if (i == 0 && opc == OST_WRITE &&
+ OBD_FAIL_CHECK(OBD_FAIL_OST_CHECKSUM_RECEIVE)) {
+ int off = local_nb[i].lnb_page_offset & ~PAGE_MASK;
+ int len = local_nb[i].lnb_len;
+ struct page *np = tgt_page_to_corrupt;
+
+ if (np) {
+ char *ptr = ll_kmap_atomic(local_nb[i].lnb_page,
+ KM_USER0);
+ char *ptr2 = page_address(np);
+
+ memcpy(ptr2 + off, ptr + off, len);
+ memcpy(ptr2 + off, "bad3", min(4, len));
+ ll_kunmap_atomic(ptr, KM_USER0);
+
+ /* LU-8376 to preserve original index for
+ * display in dump_all_bulk_pages() */
+ np->index = i;
+
+ cfs_crypto_hash_update_page(req, np, off,
+ len);
+ continue;
+ } else {
+ CERROR("%s: can't alloc page for corruption\n",
+ tgt_name(tgt));
+ }
+ }
+
+ /*
+ * The left guard number should be able to hold checksums of a
+ * whole page
+ */
+ if (t10_cksum_type && opc == OST_READ &&
+ local_nb[i].lnb_guard_disk) {
+ used = DIV_ROUND_UP(local_nb[i].lnb_len, sector_size);
+ if (used > (guard_number - used_number)) {
+ rc = -E2BIG;
+ break;
+ }
+ memcpy(guard_start + used_number,
+ local_nb[i].lnb_guards,
+ used * sizeof(*local_nb[i].lnb_guards));
+ } else {
+ rc = obd_page_dif_generate_buffer(obd_name,
+ local_nb[i].lnb_page,
+ local_nb[i].lnb_page_offset & ~PAGE_MASK,
+ local_nb[i].lnb_len, guard_start + used_number,
+ guard_number - used_number, &used, sector_size,
+ fn);
+ if (rc)
+ break;
+ }
+
+ LASSERT(used <= MAX_GUARD_NUMBER);
+ /* If disk support T10PI checksum, copy guards to local_nb */
+ if (t10_cksum_type && opc == OST_WRITE) {
+ local_nb[i].lnb_guard_rpc = 1;
+ memcpy(local_nb[i].lnb_guards,
+ guard_start + used_number,
+ used * sizeof(*local_nb[i].lnb_guards));
+ }
+
+ used_number += used;
+ if (used_number == guard_number) {
+ cfs_crypto_hash_update_page(req, __page, 0,
+ used_number * sizeof(*guard_start));
+ used_number = 0;
+ }
+
+ /* corrupt the data after we compute the checksum, to
+ * simulate an OST->client data error */
+ if (unlikely(i == 0 && opc == OST_READ &&
+ OBD_FAIL_CHECK(OBD_FAIL_OST_CHECKSUM_SEND))) {
+ int off = local_nb[i].lnb_page_offset & ~PAGE_MASK;
+ int len = local_nb[i].lnb_len;
struct page *np = tgt_page_to_corrupt;
- char *ptr = kmap(desc->bd_iov[i].kiov_page) + off;
if (np) {
- char *ptr2 = kmap(np) + off;
+ char *ptr = ll_kmap_atomic(local_nb[i].lnb_page,
+ KM_USER0);
+ char *ptr2 = page_address(np);
+
+ memcpy(ptr2 + off, ptr + off, len);
+ memcpy(ptr2 + off, "bad4", min(4, len));
+ ll_kunmap_atomic(ptr, KM_USER0);
- memcpy(ptr2, ptr, len);
- memcpy(ptr2, "bad4", min(4, len));
- kunmap(np);
- desc->bd_iov[i].kiov_page = np;
+ /* LU-8376 to preserve original index for
+ * display in dump_all_bulk_pages() */
+ np->index = i;
+
+ cfs_crypto_hash_update_page(req, np, off,
+ len);
+ continue;
} else {
CERROR("%s: can't alloc page for corruption\n",
tgt_name(tgt));
}
}
}
+ kunmap(__page);
+ if (rc)
+ GOTO(out, rc);
+
+ if (used_number != 0)
+ cfs_crypto_hash_update_page(req, __page, 0,
+ used_number * sizeof(*guard_start));
+
+ bufsize = sizeof(cksum);
+ rc = cfs_crypto_hash_final(req, (unsigned char *)&cksum, &bufsize);
+
+ if (rc == 0)
+ *check_sum = cksum;
+out:
+ __free_page(__page);
+ return rc;
+}
- bufsize = 4;
- err = cfs_crypto_hash_final(hdesc, (unsigned char *)&cksum, &bufsize);
- if (err)
- cfs_crypto_hash_final(hdesc, NULL, NULL);
+static int tgt_checksum_niobuf_rw(struct lu_target *tgt,
+ enum cksum_types cksum_type,
+ struct niobuf_local *local_nb,
+ int npages, int opc, u32 *check_sum)
+{
+ obd_dif_csum_fn *fn = NULL;
+ int sector_size = 0;
+ int rc;
+
+ ENTRY;
+ obd_t10_cksum2dif(cksum_type, &fn, §or_size);
- return cksum;
+ if (fn)
+ rc = tgt_checksum_niobuf_t10pi(tgt, local_nb, npages,
+ opc, fn, sector_size,
+ check_sum);
+ else
+ rc = tgt_checksum_niobuf(tgt, local_nb, npages, opc,
+ cksum_type, check_sum);
+ RETURN(rc);
}
int tgt_brw_read(struct tgt_session_info *tsi)
struct ost_body *body, *repbody;
struct l_wait_info lwi;
struct lustre_handle lockh = { 0 };
- int niocount, npages, nob = 0, rc, i;
- int no_reply = 0;
+ int npages, nob = 0, rc, i, no_reply = 0,
+ npages_read;
struct tgt_thread_big_cache *tbc = req->rq_svc_thread->t_data;
+ const char *obd_name = exp->exp_obd->obd_name;
ENTRY;
- if (ptlrpc_req2svc(req)->srv_req_portal != OST_IO_PORTAL) {
+ if (ptlrpc_req2svc(req)->srv_req_portal != OST_IO_PORTAL &&
+ ptlrpc_req2svc(req)->srv_req_portal != MDS_IO_PORTAL) {
CERROR("%s: deny read request from %s to portal %u\n",
tgt_name(tsi->tsi_tgt),
obd_export_nid2str(req->rq_export),
/* Check if there is eviction in progress, and if so, wait for it to
* finish */
- if (unlikely(cfs_atomic_read(&exp->exp_obd->obd_evict_inprogress))) {
+ if (unlikely(atomic_read(&exp->exp_obd->obd_evict_inprogress))) {
/* We do not care how long it takes */
lwi = LWI_INTR(NULL, NULL);
rc = l_wait_event(exp->exp_obd->obd_evict_inprogress_waitq,
- !cfs_atomic_read(&exp->exp_obd->obd_evict_inprogress),
+ !atomic_read(&exp->exp_obd->obd_evict_inprogress),
&lwi);
}
ioo = req_capsule_client_get(tsi->tsi_pill, &RMF_OBD_IOOBJ);
LASSERT(ioo != NULL); /* must exists after tgt_ost_body_unpack */
- niocount = ioo->ioo_bufcnt;
remote_nb = req_capsule_client_get(&req->rq_pill, &RMF_NIOBUF_REMOTE);
LASSERT(remote_nb != NULL); /* must exists after tgt_ost_body_unpack */
local_nb = tbc->local;
- rc = tgt_brw_lock(exp->exp_obd->obd_namespace, &tsi->tsi_resid, ioo,
- remote_nb, &lockh, LCK_PR);
+ rc = tgt_brw_lock(tsi->tsi_env, exp, &tsi->tsi_resid, ioo, remote_nb,
+ &lockh, LCK_PR);
if (rc != 0)
RETURN(rc);
* If getting the lock took more time than
* client was willing to wait, drop it. b=11330
*/
- if (cfs_time_current_sec() > req->rq_deadline ||
+ if (ktime_get_real_seconds() > req->rq_deadline ||
OBD_FAIL_CHECK(OBD_FAIL_OST_DROP_REQ)) {
no_reply = 1;
- CERROR("Dropping timed-out read from %s because locking"
- "object "DOSTID" took %ld seconds (limit was %ld).\n",
+ CERROR("Dropping timed-out read from %s because locking object " DOSTID " took %lld seconds (limit was %lld).\n",
libcfs_id2str(req->rq_peer), POSTID(&ioo->ioo_oid),
- cfs_time_current_sec() - req->rq_arrival_time.tv_sec,
+ ktime_get_real_seconds() - req->rq_arrival_time.tv_sec,
req->rq_deadline - req->rq_arrival_time.tv_sec);
GOTO(out_lock, rc = -ETIMEDOUT);
}
npages = PTLRPC_MAX_BRW_PAGES;
rc = obd_preprw(tsi->tsi_env, OBD_BRW_READ, exp, &repbody->oa, 1,
- ioo, remote_nb, &npages, local_nb, NULL, BYPASS_CAPA);
+ ioo, remote_nb, &npages, local_nb);
if (rc != 0)
GOTO(out_lock, rc);
- desc = ptlrpc_prep_bulk_exp(req, npages, ioobj_max_brw_get(ioo),
- BULK_PUT_SOURCE, OST_BULK_PORTAL);
- if (desc == NULL)
- GOTO(out_commitrw, rc = -ENOMEM);
+ if (body->oa.o_flags & OBD_FL_SHORT_IO) {
+ desc = NULL;
+ } else {
+ desc = ptlrpc_prep_bulk_exp(req, npages, ioobj_max_brw_get(ioo),
+ PTLRPC_BULK_PUT_SOURCE |
+ PTLRPC_BULK_BUF_KIOV,
+ OST_BULK_PORTAL,
+ &ptlrpc_bulk_kiov_nopin_ops);
+ if (desc == NULL)
+ GOTO(out_commitrw, rc = -ENOMEM);
+ }
nob = 0;
+ npages_read = npages;
for (i = 0; i < npages; i++) {
- int page_rc = local_nb[i].rc;
+ int page_rc = local_nb[i].lnb_rc;
if (page_rc < 0) {
rc = page_rc;
+ npages_read = i;
break;
}
nob += page_rc;
- if (page_rc != 0) { /* some data! */
- LASSERT(local_nb[i].page != NULL);
- ptlrpc_prep_bulk_page_nopin(desc, local_nb[i].page,
- local_nb[i].lnb_page_offset,
- page_rc);
+ if (page_rc != 0 && desc != NULL) { /* some data! */
+ LASSERT(local_nb[i].lnb_page != NULL);
+ desc->bd_frag_ops->add_kiov_frag
+ (desc, local_nb[i].lnb_page,
+ local_nb[i].lnb_page_offset & ~PAGE_MASK,
+ page_rc);
}
- if (page_rc != local_nb[i].len) { /* short read */
+ if (page_rc != local_nb[i].lnb_len) { /* short read */
+ local_nb[i].lnb_len = page_rc;
+ npages_read = i + (page_rc != 0 ? 1 : 0);
/* All subsequent pages should be 0 */
while (++i < npages)
- LASSERT(local_nb[i].rc == 0);
+ LASSERT(local_nb[i].lnb_rc == 0);
break;
}
}
+ if (OBD_FAIL_CHECK(OBD_FAIL_OST_READ_SIZE) &&
+ nob != cfs_fail_val)
+ rc = -E2BIG;
if (body->oa.o_valid & OBD_MD_FLCKSUM) {
- cksum_type_t cksum_type =
- cksum_type_unpack(body->oa.o_valid & OBD_MD_FLFLAGS ?
- body->oa.o_flags : 0);
- repbody->oa.o_flags = cksum_type_pack(cksum_type);
+ u32 flag = body->oa.o_valid & OBD_MD_FLFLAGS ?
+ body->oa.o_flags : 0;
+ enum cksum_types cksum_type = obd_cksum_type_unpack(flag);
+
+ repbody->oa.o_flags = obd_cksum_type_pack(obd_name,
+ cksum_type);
repbody->oa.o_valid = OBD_MD_FLCKSUM | OBD_MD_FLFLAGS;
- repbody->oa.o_cksum = tgt_checksum_bulk(tsi->tsi_tgt, desc,
- OST_READ, cksum_type);
+
+ rc = tgt_checksum_niobuf_rw(tsi->tsi_tgt, cksum_type,
+ local_nb, npages_read, OST_READ,
+ &repbody->oa.o_cksum);
+ if (rc < 0)
+ GOTO(out_commitrw, rc);
CDEBUG(D_PAGE, "checksum at read origin: %x\n",
repbody->oa.o_cksum);
+
+ /* if a resend it could be for a cksum error, so check Server
+ * cksum with returned Client cksum (this should even cover
+ * zero-cksum case) */
+ if ((body->oa.o_valid & OBD_MD_FLFLAGS) &&
+ (body->oa.o_flags & OBD_FL_RECOV_RESEND))
+ check_read_checksum(local_nb, npages_read, exp,
+ &body->oa, &req->rq_peer,
+ body->oa.o_cksum,
+ repbody->oa.o_cksum, cksum_type);
} else {
repbody->oa.o_valid = 0;
}
/* Check if client was evicted while we were doing i/o before touching
* network */
- if (likely(rc == 0 &&
- !CFS_FAIL_PRECHECK(OBD_FAIL_PTLRPC_CLIENT_BULK_CB2))) {
- rc = target_bulk_io(exp, desc, &lwi);
+ if (rc == 0) {
+ if (body->oa.o_flags & OBD_FL_SHORT_IO) {
+ unsigned char *short_io_buf;
+ int short_io_size;
+
+ short_io_buf = req_capsule_server_get(&req->rq_pill,
+ &RMF_SHORT_IO);
+ short_io_size = req_capsule_get_size(&req->rq_pill,
+ &RMF_SHORT_IO,
+ RCL_SERVER);
+ rc = tgt_pages2shortio(local_nb, npages_read,
+ short_io_buf, short_io_size);
+ if (rc >= 0)
+ req_capsule_shrink(&req->rq_pill,
+ &RMF_SHORT_IO, rc,
+ RCL_SERVER);
+ rc = rc > 0 ? 0 : rc;
+ } else if (!CFS_FAIL_PRECHECK(OBD_FAIL_PTLRPC_CLIENT_BULK_CB2)) {
+ rc = target_bulk_io(exp, desc, &lwi);
+ }
no_reply = rc != 0;
+ } else {
+ if (body->oa.o_flags & OBD_FL_SHORT_IO)
+ req_capsule_shrink(&req->rq_pill, &RMF_SHORT_IO, 0,
+ RCL_SERVER);
}
out_commitrw:
/* Must commit after prep above in all cases */
- rc = obd_commitrw(tsi->tsi_env, OBD_BRW_READ, exp,
- &repbody->oa, 1, ioo, remote_nb, npages, local_nb,
- NULL, rc);
- if (rc == 0)
- tgt_drop_id(exp, &repbody->oa);
+ rc = obd_commitrw(tsi->tsi_env, OBD_BRW_READ, exp, &repbody->oa, 1, ioo,
+ remote_nb, npages, local_nb, rc);
out_lock:
tgt_brw_unlock(ioo, remote_nb, &lockh, LCK_PR);
if (desc && !CFS_FAIL_PRECHECK(OBD_FAIL_PTLRPC_CLIENT_BULK_CB2))
- ptlrpc_free_bulk_nopin(desc);
+ ptlrpc_free_bulk(desc);
LASSERT(rc <= 0);
if (rc == 0) {
ptlrpc_req_drop_rs(req);
LCONSOLE_WARN("%s: Bulk IO read error with %s (at %s), "
"client will retry: rc %d\n",
- exp->exp_obd->obd_name,
+ obd_name,
obd_uuid2str(&exp->exp_client_uuid),
obd_export_nid2str(exp), rc);
}
/* send a bulk after reply to simulate a network delay or reordering
- * by a router */
- if (unlikely(CFS_FAIL_PRECHECK(OBD_FAIL_PTLRPC_CLIENT_BULK_CB2))) {
+ * by a router - Note that !desc implies short io, so there is no bulk
+ * to reorder. */
+ if (unlikely(CFS_FAIL_PRECHECK(OBD_FAIL_PTLRPC_CLIENT_BULK_CB2)) &&
+ desc) {
wait_queue_head_t waitq;
struct l_wait_info lwi1;
lwi1 = LWI_TIMEOUT_INTR(cfs_time_seconds(3), NULL, NULL, NULL);
l_wait_event(waitq, 0, &lwi1);
target_bulk_io(exp, desc, &lwi);
- ptlrpc_free_bulk_nopin(desc);
+ ptlrpc_free_bulk(desc);
}
RETURN(rc);
}
EXPORT_SYMBOL(tgt_brw_read);
+static int tgt_shortio2pages(struct niobuf_local *local, int npages,
+ unsigned char *buf, int size)
+{
+ int i, off, len;
+ char *ptr;
+
+ for (i = 0; i < npages; i++) {
+ off = local[i].lnb_page_offset & ~PAGE_MASK;
+ len = local[i].lnb_len;
+
+ if (len == 0)
+ continue;
+
+ CDEBUG(D_PAGE, "index %d offset = %d len = %d left = %d\n",
+ i, off, len, size);
+ ptr = ll_kmap_atomic(local[i].lnb_page, KM_USER0);
+ if (ptr == NULL)
+ return -EINVAL;
+ memcpy(ptr + off, buf, len < size ? len : size);
+ ll_kunmap_atomic(ptr, KM_USER0);
+ buf += len;
+ size -= len;
+ }
+ return 0;
+}
+
static void tgt_warn_on_cksum(struct ptlrpc_request *req,
struct ptlrpc_bulk_desc *desc,
struct niobuf_local *local_nb, int npages,
- obd_count client_cksum, obd_count server_cksum,
+ u32 client_cksum, u32 server_cksum,
bool mmap)
{
struct obd_export *exp = req->rq_export;
struct ost_body *body;
- char *router;
- char *via;
+ char *router = "";
+ char *via = "";
body = req_capsule_client_get(&req->rq_pill, &RMF_OST_BODY);
LASSERT(body != NULL);
- if (req->rq_peer.nid == desc->bd_sender) {
- via = router = "";
- } else {
+ if (desc && req->rq_peer.nid != desc->bd_sender) {
via = " via ";
router = libcfs_nid2str(desc->bd_sender);
}
+ if (exp->exp_obd->obd_checksum_dump)
+ dump_all_bulk_pages(&body->oa, npages, local_nb, server_cksum,
+ client_cksum);
+
if (mmap) {
CDEBUG_LIMIT(D_INFO, "client csum %x, server csum %x\n",
client_cksum, server_cksum);
return;
}
- LCONSOLE_ERROR_MSG(0x168, "BAD WRITE CHECKSUM: %s from %s%s%s inode "
- DFID" object "DOSTID" extent ["LPU64"-"LPU64
+ LCONSOLE_ERROR_MSG(0x168, "%s: BAD WRITE CHECKSUM: from %s%s%s inode "
+ DFID" object "DOSTID" extent [%llu-%llu"
"]: client csum %x, server csum %x\n",
exp->exp_obd->obd_name, libcfs_id2str(req->rq_peer),
via, router,
POSTID(&body->oa.o_oi),
local_nb[0].lnb_file_offset,
local_nb[npages-1].lnb_file_offset +
- local_nb[npages-1].len - 1,
+ local_nb[npages - 1].lnb_len - 1,
client_cksum, server_cksum);
}
__u32 *rcs;
int objcount, niocount, npages;
int rc, i, j;
- cksum_type_t cksum_type = OBD_CKSUM_CRC32;
+ enum cksum_types cksum_type = OBD_CKSUM_CRC32;
bool no_reply = false, mmap;
struct tgt_thread_big_cache *tbc = req->rq_svc_thread->t_data;
+ bool wait_sync = false;
+ const char *obd_name = exp->exp_obd->obd_name;
ENTRY;
- if (ptlrpc_req2svc(req)->srv_req_portal != OST_IO_PORTAL) {
+ if (ptlrpc_req2svc(req)->srv_req_portal != OST_IO_PORTAL &&
+ ptlrpc_req2svc(req)->srv_req_portal != MDS_IO_PORTAL) {
CERROR("%s: deny write request from %s to portal %u\n",
tgt_name(tsi->tsi_tgt),
obd_export_nid2str(req->rq_export),
sizeof(*remote_nb))
RETURN(err_serious(-EPROTO));
- if ((remote_nb[0].flags & OBD_BRW_MEMALLOC) &&
- (exp->exp_connection->c_peer.nid == exp->exp_connection->c_self))
+ if ((remote_nb[0].rnb_flags & OBD_BRW_MEMALLOC) &&
+ ptlrpc_connection_is_local(exp->exp_connection))
memory_pressure_set();
req_capsule_set_size(&req->rq_pill, &RMF_RCS, RCL_SERVER,
local_nb = tbc->local;
- rc = tgt_brw_lock(exp->exp_obd->obd_namespace, &tsi->tsi_resid, ioo,
- remote_nb, &lockh, LCK_PW);
+ rc = tgt_brw_lock(tsi->tsi_env, exp, &tsi->tsi_resid, ioo, remote_nb,
+ &lockh, LCK_PW);
if (rc != 0)
GOTO(out, rc);
* If getting the lock took more time than
* client was willing to wait, drop it. b=11330
*/
- if (cfs_time_current_sec() > req->rq_deadline ||
+ if (ktime_get_real_seconds() > req->rq_deadline ||
OBD_FAIL_CHECK(OBD_FAIL_OST_DROP_REQ)) {
no_reply = true;
- CERROR("%s: Dropping timed-out write from %s because locking "
- "object "DOSTID" took %ld seconds (limit was %ld).\n",
+ CERROR("%s: Dropping timed-out write from %s because locking object " DOSTID " took %lld seconds (limit was %lld).\n",
tgt_name(tsi->tsi_tgt), libcfs_id2str(req->rq_peer),
POSTID(&ioo->ioo_oid),
- cfs_time_current_sec() - req->rq_arrival_time.tv_sec,
+ ktime_get_real_seconds() - req->rq_arrival_time.tv_sec,
req->rq_deadline - req->rq_arrival_time.tv_sec);
GOTO(out_lock, rc = -ETIMEDOUT);
}
npages = PTLRPC_MAX_BRW_PAGES;
rc = obd_preprw(tsi->tsi_env, OBD_BRW_WRITE, exp, &repbody->oa,
- objcount, ioo, remote_nb, &npages, local_nb, NULL,
- BYPASS_CAPA);
+ objcount, ioo, remote_nb, &npages, local_nb);
if (rc < 0)
GOTO(out_lock, rc);
+ if (body->oa.o_flags & OBD_FL_SHORT_IO) {
+ int short_io_size;
+ unsigned char *short_io_buf;
+
+ short_io_size = req_capsule_get_size(&req->rq_pill,
+ &RMF_SHORT_IO,
+ RCL_CLIENT);
+ short_io_buf = req_capsule_client_get(&req->rq_pill,
+ &RMF_SHORT_IO);
+ CDEBUG(D_INFO, "Client use short io for data transfer,"
+ " size = %d\n", short_io_size);
+
+ /* Copy short io buf to pages */
+ rc = tgt_shortio2pages(local_nb, npages, short_io_buf,
+ short_io_size);
+ desc = NULL;
+ } else {
+ desc = ptlrpc_prep_bulk_exp(req, npages, ioobj_max_brw_get(ioo),
+ PTLRPC_BULK_GET_SINK |
+ PTLRPC_BULK_BUF_KIOV,
+ OST_BULK_PORTAL,
+ &ptlrpc_bulk_kiov_nopin_ops);
+ if (desc == NULL)
+ GOTO(skip_transfer, rc = -ENOMEM);
+
+ /* NB Having prepped, we must commit... */
+ for (i = 0; i < npages; i++)
+ desc->bd_frag_ops->add_kiov_frag(desc,
+ local_nb[i].lnb_page,
+ local_nb[i].lnb_page_offset & ~PAGE_MASK,
+ local_nb[i].lnb_len);
+
+ rc = sptlrpc_svc_prep_bulk(req, desc);
+ if (rc != 0)
+ GOTO(skip_transfer, rc);
- desc = ptlrpc_prep_bulk_exp(req, npages, ioobj_max_brw_get(ioo),
- BULK_GET_SINK, OST_BULK_PORTAL);
- if (desc == NULL)
- GOTO(skip_transfer, rc = -ENOMEM);
-
- /* NB Having prepped, we must commit... */
- for (i = 0; i < npages; i++)
- ptlrpc_prep_bulk_page_nopin(desc, local_nb[i].page,
- local_nb[i].lnb_page_offset,
- local_nb[i].len);
-
- rc = sptlrpc_svc_prep_bulk(req, desc);
- if (rc != 0)
- GOTO(skip_transfer, rc);
+ rc = target_bulk_io(exp, desc, &lwi);
+ }
- rc = target_bulk_io(exp, desc, &lwi);
no_reply = rc != 0;
skip_transfer:
static int cksum_counter;
if (body->oa.o_valid & OBD_MD_FLFLAGS)
- cksum_type = cksum_type_unpack(body->oa.o_flags);
+ cksum_type = obd_cksum_type_unpack(body->oa.o_flags);
repbody->oa.o_valid |= OBD_MD_FLCKSUM | OBD_MD_FLFLAGS;
repbody->oa.o_flags &= ~OBD_FL_CKSUM_ALL;
- repbody->oa.o_flags |= cksum_type_pack(cksum_type);
- repbody->oa.o_cksum = tgt_checksum_bulk(tsi->tsi_tgt, desc,
- OST_WRITE, cksum_type);
+ repbody->oa.o_flags |= obd_cksum_type_pack(obd_name,
+ cksum_type);
+
+ rc = tgt_checksum_niobuf_rw(tsi->tsi_tgt, cksum_type,
+ local_nb, npages, OST_WRITE,
+ &repbody->oa.o_cksum);
+ if (rc < 0)
+ GOTO(out_commitrw, rc);
+
cksum_counter++;
if (unlikely(body->oa.o_cksum != repbody->oa.o_cksum)) {
}
}
+out_commitrw:
/* Must commit after prep above in all cases */
rc = obd_commitrw(tsi->tsi_env, OBD_BRW_WRITE, exp, &repbody->oa,
- objcount, ioo, remote_nb, npages, local_nb, NULL,
- rc);
+ objcount, ioo, remote_nb, npages, local_nb, rc);
if (rc == -ENOTCONN)
/* quota acquire process has been given up because
* either the client has been evicted or the client
* has timed out the request already */
no_reply = true;
+ for (i = 0; i < niocount; i++) {
+ if (!(local_nb[i].lnb_flags & OBD_BRW_ASYNC)) {
+ wait_sync = true;
+ break;
+ }
+ }
/*
* Disable sending mtime back to the client. If the client locked the
* whole object, then it has already updated the mtime on its side,
/* set per-requested niobuf return codes */
for (i = j = 0; i < niocount; i++) {
- int len = remote_nb[i].len;
+ int len = remote_nb[i].rnb_len;
nob += len;
rcs[i] = 0;
do {
LASSERT(j < npages);
- if (local_nb[j].rc < 0)
- rcs[i] = local_nb[j].rc;
- len -= local_nb[j].len;
+ if (local_nb[j].lnb_rc < 0)
+ rcs[i] = local_nb[j].lnb_rc;
+ len -= local_nb[j].lnb_len;
j++;
} while (len > 0);
LASSERT(len == 0);
}
LASSERT(j == npages);
ptlrpc_lprocfs_brw(req, nob);
-
- tgt_drop_id(exp, &repbody->oa);
}
out_lock:
tgt_brw_unlock(ioo, remote_nb, &lockh, LCK_PW);
if (desc)
- ptlrpc_free_bulk_nopin(desc);
+ ptlrpc_free_bulk(desc);
out:
- if (no_reply) {
+ if (unlikely(no_reply || (exp->exp_obd->obd_no_transno && wait_sync))) {
req->rq_no_reply = 1;
/* reply out callback would free */
ptlrpc_req_drop_rs(req);
- LCONSOLE_WARN("%s: Bulk IO write error with %s (at %s), "
- "client will retry: rc %d\n",
- exp->exp_obd->obd_name,
- obd_uuid2str(&exp->exp_client_uuid),
- obd_export_nid2str(exp), rc);
+ if (!exp->exp_obd->obd_no_transno)
+ LCONSOLE_WARN("%s: Bulk IO write error with %s (at %s),"
+ " client will retry: rc = %d\n",
+ obd_name,
+ obd_uuid2str(&exp->exp_client_uuid),
+ obd_export_nid2str(exp), rc);
}
memory_pressure_clr();
RETURN(rc);
}
EXPORT_SYMBOL(tgt_brw_write);
+
+/* Check if request can be reconstructed from saved reply data
+ * A copy of the reply data is returned in @trd if the pointer is not NULL
+ */
+bool req_can_reconstruct(struct ptlrpc_request *req,
+ struct tg_reply_data *trd)
+{
+ struct tg_export_data *ted = &req->rq_export->exp_target_data;
+ struct lsd_client_data *lcd = ted->ted_lcd;
+ bool found;
+
+ if (tgt_is_multimodrpcs_client(req->rq_export))
+ return tgt_lookup_reply(req, trd);
+
+ mutex_lock(&ted->ted_lcd_lock);
+ found = req->rq_xid == lcd->lcd_last_xid ||
+ req->rq_xid == lcd->lcd_last_close_xid;
+
+ if (found && trd != NULL) {
+ if (lustre_msg_get_opc(req->rq_reqmsg) == MDS_CLOSE) {
+ trd->trd_reply.lrd_xid = lcd->lcd_last_close_xid;
+ trd->trd_reply.lrd_transno =
+ lcd->lcd_last_close_transno;
+ trd->trd_reply.lrd_result = lcd->lcd_last_close_result;
+ } else {
+ trd->trd_reply.lrd_xid = lcd->lcd_last_xid;
+ trd->trd_reply.lrd_transno = lcd->lcd_last_transno;
+ trd->trd_reply.lrd_result = lcd->lcd_last_result;
+ trd->trd_reply.lrd_data = lcd->lcd_last_data;
+ trd->trd_pre_versions[0] = lcd->lcd_pre_versions[0];
+ trd->trd_pre_versions[1] = lcd->lcd_pre_versions[1];
+ trd->trd_pre_versions[2] = lcd->lcd_pre_versions[2];
+ trd->trd_pre_versions[3] = lcd->lcd_pre_versions[3];
+ }
+ }
+ mutex_unlock(&ted->ted_lcd_lock);
+
+ return found;
+}
+EXPORT_SYMBOL(req_can_reconstruct);
+