4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 only,
8 * as published by the Free Software Foundation.
10 * This program is distributed in the hope that it will be useful, but
11 * WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * General Public License version 2 for more details (a copy is included
14 * in the LICENSE file that accompanied this code).
16 * You should have received a copy of the GNU General Public License
17 * version 2 along with this program; If not, see
18 * http://www.gnu.org/licenses/gpl-2.0.html
23 * Copyright (c) 2002, 2010, Oracle and/or its affiliates. All rights reserved.
24 * Use is subject to license terms.
26 * Copyright (c) 2011, 2017, Intel Corporation.
29 * This file is part of Lustre, http://www.lustre.org/
30 * Lustre is a trademark of Sun Microsystems, Inc.
32 * lustre/ptlrpc/pack_generic.c
34 * (Un)packing of OST requests
36 * Author: Peter J. Braam <braam@clusterfs.com>
37 * Author: Phil Schwan <phil@clusterfs.com>
38 * Author: Eric Barton <eeb@clusterfs.com>
41 #define DEBUG_SUBSYSTEM S_RPC
43 #include <libcfs/libcfs.h>
45 #include <llog_swab.h>
46 #include <lustre_net.h>
47 #include <lustre_swab.h>
48 #include <obd_cksum.h>
49 #include <obd_class.h>
50 #include <obd_support.h>
51 #include <obj_update.h>
53 #include "ptlrpc_internal.h"
55 static inline __u32 lustre_msg_hdr_size_v2(__u32 count)
57 return cfs_size_round(offsetof(struct lustre_msg_v2,
61 __u32 lustre_msg_hdr_size(__u32 magic, __u32 count)
66 case LUSTRE_MSG_MAGIC_V2:
67 return lustre_msg_hdr_size_v2(count);
69 LASSERTF(0, "incorrect message magic: %08x\n", magic);
74 void ptlrpc_buf_set_swabbed(struct ptlrpc_request *req, const int inout,
78 lustre_set_req_swabbed(req, index);
80 lustre_set_rep_swabbed(req, index);
83 bool ptlrpc_buf_need_swab(struct ptlrpc_request *req, const int inout,
87 return (ptlrpc_req_need_swab(req) &&
88 !lustre_req_swabbed(req, index));
90 return (ptlrpc_rep_need_swab(req) && !lustre_rep_swabbed(req, index));
93 static inline int lustre_msg_check_version_v2(struct lustre_msg_v2 *msg,
94 enum lustre_msg_version version)
96 enum lustre_msg_version ver = lustre_msg_get_version(msg);
98 return (ver & LUSTRE_VERSION_MASK) != version;
101 int lustre_msg_check_version(struct lustre_msg *msg,
102 enum lustre_msg_version version)
104 #define LUSTRE_MSG_MAGIC_V1 0x0BD00BD0
105 switch (msg->lm_magic) {
106 case LUSTRE_MSG_MAGIC_V1:
107 CERROR("msg v1 not supported - please upgrade you system\n");
109 case LUSTRE_MSG_MAGIC_V2:
110 return lustre_msg_check_version_v2(msg, version);
112 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
115 #undef LUSTRE_MSG_MAGIC_V1
118 /* early reply size */
119 __u32 lustre_msg_early_size()
121 __u32 pblen = sizeof(struct ptlrpc_body);
123 return lustre_msg_size(LUSTRE_MSG_MAGIC_V2, 1, &pblen);
125 EXPORT_SYMBOL(lustre_msg_early_size);
127 __u32 lustre_msg_size_v2(int count, __u32 *lengths)
133 size = lustre_msg_hdr_size_v2(count);
134 for (i = 0; i < count; i++)
135 size += cfs_size_round(lengths[i]);
139 EXPORT_SYMBOL(lustre_msg_size_v2);
142 * This returns the size of the buffer that is required to hold a lustre_msg
143 * with the given sub-buffer lengths.
144 * NOTE: this should only be used for NEW requests, and should always be
145 * in the form of a v2 request. If this is a connection to a v1
146 * target then the first buffer will be stripped because the ptlrpc
147 * data is part of the lustre_msg_v1 header. b=14043
149 __u32 lustre_msg_size(__u32 magic, int count, __u32 *lens)
151 __u32 size[] = { sizeof(struct ptlrpc_body) };
159 LASSERT(lens[MSG_PTLRPC_BODY_OFF] >= sizeof(struct ptlrpc_body_v2));
162 case LUSTRE_MSG_MAGIC_V2:
163 return lustre_msg_size_v2(count, lens);
165 LASSERTF(0, "incorrect message magic: %08x\n", magic);
171 * This is used to determine the size of a buffer that was already packed
172 * and will correctly handle the different message formats.
174 __u32 lustre_packed_msg_size(struct lustre_msg *msg)
176 switch (msg->lm_magic) {
177 case LUSTRE_MSG_MAGIC_V2:
178 return lustre_msg_size_v2(msg->lm_bufcount, msg->lm_buflens);
180 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
184 EXPORT_SYMBOL(lustre_packed_msg_size);
186 void lustre_init_msg_v2(struct lustre_msg_v2 *msg, int count, __u32 *lens,
194 msg->lm_bufcount = count;
195 /* XXX: lm_secflvr uninitialized here */
196 msg->lm_magic = LUSTRE_MSG_MAGIC_V2;
198 for (i = 0; i < count; i++)
199 msg->lm_buflens[i] = lens[i];
204 ptr = (char *)msg + lustre_msg_hdr_size_v2(count);
205 for (i = 0; i < count; i++) {
209 memcpy(ptr, tmp, lens[i]);
210 ptr += cfs_size_round(lens[i]);
213 EXPORT_SYMBOL(lustre_init_msg_v2);
215 static int lustre_pack_request_v2(struct ptlrpc_request *req,
216 int count, __u32 *lens, char **bufs)
220 reqlen = lustre_msg_size_v2(count, lens);
222 rc = sptlrpc_cli_alloc_reqbuf(req, reqlen);
226 req->rq_reqlen = reqlen;
228 lustre_init_msg_v2(req->rq_reqmsg, count, lens, bufs);
229 lustre_msg_add_version(req->rq_reqmsg, PTLRPC_MSG_VERSION);
233 int lustre_pack_request(struct ptlrpc_request *req, __u32 magic, int count,
234 __u32 *lens, char **bufs)
236 __u32 size[] = { sizeof(struct ptlrpc_body) };
244 LASSERT(lens[MSG_PTLRPC_BODY_OFF] == sizeof(struct ptlrpc_body));
246 /* only use new format, we don't need to be compatible with 1.4 */
247 magic = LUSTRE_MSG_MAGIC_V2;
250 case LUSTRE_MSG_MAGIC_V2:
251 return lustre_pack_request_v2(req, count, lens, bufs);
253 LASSERTF(0, "incorrect message magic: %08x\n", magic);
259 struct list_head ptlrpc_rs_debug_lru =
260 LIST_HEAD_INIT(ptlrpc_rs_debug_lru);
261 spinlock_t ptlrpc_rs_debug_lock;
263 #define PTLRPC_RS_DEBUG_LRU_ADD(rs) \
265 spin_lock(&ptlrpc_rs_debug_lock); \
266 list_add_tail(&(rs)->rs_debug_list, &ptlrpc_rs_debug_lru); \
267 spin_unlock(&ptlrpc_rs_debug_lock); \
270 #define PTLRPC_RS_DEBUG_LRU_DEL(rs) \
272 spin_lock(&ptlrpc_rs_debug_lock); \
273 list_del(&(rs)->rs_debug_list); \
274 spin_unlock(&ptlrpc_rs_debug_lock); \
277 # define PTLRPC_RS_DEBUG_LRU_ADD(rs) do {} while(0)
278 # define PTLRPC_RS_DEBUG_LRU_DEL(rs) do {} while(0)
281 struct ptlrpc_reply_state *
282 lustre_get_emerg_rs(struct ptlrpc_service_part *svcpt)
284 struct ptlrpc_reply_state *rs = NULL;
286 spin_lock(&svcpt->scp_rep_lock);
288 /* See if we have anything in a pool, and wait if nothing */
289 while (list_empty(&svcpt->scp_rep_idle)) {
290 struct l_wait_info lwi;
293 spin_unlock(&svcpt->scp_rep_lock);
294 /* If we cannot get anything for some long time, we better
295 * bail out instead of waiting infinitely */
296 lwi = LWI_TIMEOUT(cfs_time_seconds(10), NULL, NULL);
297 rc = l_wait_event(svcpt->scp_rep_waitq,
298 !list_empty(&svcpt->scp_rep_idle), &lwi);
301 spin_lock(&svcpt->scp_rep_lock);
304 rs = list_entry(svcpt->scp_rep_idle.next,
305 struct ptlrpc_reply_state, rs_list);
306 list_del(&rs->rs_list);
308 spin_unlock(&svcpt->scp_rep_lock);
310 memset(rs, 0, svcpt->scp_service->srv_max_reply_size);
311 rs->rs_size = svcpt->scp_service->srv_max_reply_size;
312 rs->rs_svcpt = svcpt;
318 void lustre_put_emerg_rs(struct ptlrpc_reply_state *rs)
320 struct ptlrpc_service_part *svcpt = rs->rs_svcpt;
322 spin_lock(&svcpt->scp_rep_lock);
323 list_add(&rs->rs_list, &svcpt->scp_rep_idle);
324 spin_unlock(&svcpt->scp_rep_lock);
325 wake_up(&svcpt->scp_rep_waitq);
328 int lustre_pack_reply_v2(struct ptlrpc_request *req, int count,
329 __u32 *lens, char **bufs, int flags)
331 struct ptlrpc_reply_state *rs;
335 LASSERT(req->rq_reply_state == NULL);
338 if ((flags & LPRFL_EARLY_REPLY) == 0) {
339 spin_lock(&req->rq_lock);
340 req->rq_packed_final = 1;
341 spin_unlock(&req->rq_lock);
344 msg_len = lustre_msg_size_v2(count, lens);
345 rc = sptlrpc_svc_alloc_rs(req, msg_len);
349 rs = req->rq_reply_state;
350 atomic_set(&rs->rs_refcount, 1); /* 1 ref for rq_reply_state */
351 rs->rs_cb_id.cbid_fn = reply_out_callback;
352 rs->rs_cb_id.cbid_arg = rs;
353 rs->rs_svcpt = req->rq_rqbd->rqbd_svcpt;
354 INIT_LIST_HEAD(&rs->rs_exp_list);
355 INIT_LIST_HEAD(&rs->rs_obd_list);
356 INIT_LIST_HEAD(&rs->rs_list);
357 spin_lock_init(&rs->rs_lock);
359 req->rq_replen = msg_len;
360 req->rq_reply_state = rs;
361 req->rq_repmsg = rs->rs_msg;
363 lustre_init_msg_v2(rs->rs_msg, count, lens, bufs);
364 lustre_msg_add_version(rs->rs_msg, PTLRPC_MSG_VERSION);
366 PTLRPC_RS_DEBUG_LRU_ADD(rs);
370 EXPORT_SYMBOL(lustre_pack_reply_v2);
372 int lustre_pack_reply_flags(struct ptlrpc_request *req, int count, __u32 *lens,
373 char **bufs, int flags)
376 __u32 size[] = { sizeof(struct ptlrpc_body) };
384 LASSERT(lens[MSG_PTLRPC_BODY_OFF] == sizeof(struct ptlrpc_body));
386 switch (req->rq_reqmsg->lm_magic) {
387 case LUSTRE_MSG_MAGIC_V2:
388 rc = lustre_pack_reply_v2(req, count, lens, bufs, flags);
391 LASSERTF(0, "incorrect message magic: %08x\n",
392 req->rq_reqmsg->lm_magic);
396 CERROR("lustre_pack_reply failed: rc=%d size=%d\n", rc,
397 lustre_msg_size(req->rq_reqmsg->lm_magic, count, lens));
401 int lustre_pack_reply(struct ptlrpc_request *req, int count, __u32 *lens,
404 return lustre_pack_reply_flags(req, count, lens, bufs, 0);
406 EXPORT_SYMBOL(lustre_pack_reply);
408 void *lustre_msg_buf_v2(struct lustre_msg_v2 *m, __u32 n, __u32 min_size)
410 __u32 i, offset, buflen, bufcount;
413 LASSERT(m->lm_bufcount > 0);
415 bufcount = m->lm_bufcount;
416 if (unlikely(n >= bufcount)) {
417 CDEBUG(D_INFO, "msg %p buffer[%d] not present (count %d)\n",
422 buflen = m->lm_buflens[n];
423 if (unlikely(buflen < min_size)) {
424 CERROR("msg %p buffer[%d] size %d too small "
425 "(required %d, opc=%d)\n", m, n, buflen, min_size,
426 n == MSG_PTLRPC_BODY_OFF ? -1 : lustre_msg_get_opc(m));
430 offset = lustre_msg_hdr_size_v2(bufcount);
431 for (i = 0; i < n; i++)
432 offset += cfs_size_round(m->lm_buflens[i]);
434 return (char *)m + offset;
437 void *lustre_msg_buf(struct lustre_msg *m, __u32 n, __u32 min_size)
439 switch (m->lm_magic) {
440 case LUSTRE_MSG_MAGIC_V2:
441 return lustre_msg_buf_v2(m, n, min_size);
443 LASSERTF(0, "incorrect message magic: %08x (msg:%p)\n",
448 EXPORT_SYMBOL(lustre_msg_buf);
450 static int lustre_shrink_msg_v2(struct lustre_msg_v2 *msg, __u32 segment,
451 unsigned int newlen, int move_data)
453 char *tail = NULL, *newpos;
457 LASSERT(msg->lm_bufcount > segment);
458 LASSERT(msg->lm_buflens[segment] >= newlen);
460 if (msg->lm_buflens[segment] == newlen)
463 if (move_data && msg->lm_bufcount > segment + 1) {
464 tail = lustre_msg_buf_v2(msg, segment + 1, 0);
465 for (n = segment + 1; n < msg->lm_bufcount; n++)
466 tail_len += cfs_size_round(msg->lm_buflens[n]);
469 msg->lm_buflens[segment] = newlen;
471 if (tail && tail_len) {
472 newpos = lustre_msg_buf_v2(msg, segment + 1, 0);
473 LASSERT(newpos <= tail);
475 memmove(newpos, tail, tail_len);
478 return lustre_msg_size_v2(msg->lm_bufcount, msg->lm_buflens);
482 * for @msg, shrink @segment to size @newlen. if @move_data is non-zero,
483 * we also move data forward from @segment + 1.
485 * if @newlen == 0, we remove the segment completely, but we still keep the
486 * totally bufcount the same to save possible data moving. this will leave a
487 * unused segment with size 0 at the tail, but that's ok.
489 * return new msg size after shrinking.
492 * + if any buffers higher than @segment has been filled in, must call shrink
493 * with non-zero @move_data.
494 * + caller should NOT keep pointers to msg buffers which higher than @segment
497 int lustre_shrink_msg(struct lustre_msg *msg, int segment,
498 unsigned int newlen, int move_data)
500 switch (msg->lm_magic) {
501 case LUSTRE_MSG_MAGIC_V2:
502 return lustre_shrink_msg_v2(msg, segment, newlen, move_data);
504 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
507 EXPORT_SYMBOL(lustre_shrink_msg);
509 static int lustre_grow_msg_v2(struct lustre_msg_v2 *msg, __u32 segment,
512 char *tail = NULL, *newpos;
516 LASSERT(msg->lm_bufcount > segment);
517 LASSERT(msg->lm_buflens[segment] <= newlen);
519 if (msg->lm_buflens[segment] == newlen)
522 if (msg->lm_bufcount > segment + 1) {
523 tail = lustre_msg_buf_v2(msg, segment + 1, 0);
524 for (n = segment + 1; n < msg->lm_bufcount; n++)
525 tail_len += cfs_size_round(msg->lm_buflens[n]);
528 msg->lm_buflens[segment] = newlen;
530 if (tail && tail_len) {
531 newpos = lustre_msg_buf_v2(msg, segment + 1, 0);
532 memmove(newpos, tail, tail_len);
535 return lustre_msg_size_v2(msg->lm_bufcount, msg->lm_buflens);
539 * for @msg, grow @segment to size @newlen.
540 * Always move higher buffer forward.
542 * return new msg size after growing.
545 * - caller must make sure there is enough space in allocated message buffer
546 * - caller should NOT keep pointers to msg buffers which higher than @segment
549 int lustre_grow_msg(struct lustre_msg *msg, int segment, unsigned int newlen)
551 switch (msg->lm_magic) {
552 case LUSTRE_MSG_MAGIC_V2:
553 return lustre_grow_msg_v2(msg, segment, newlen);
555 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
558 EXPORT_SYMBOL(lustre_grow_msg);
560 void lustre_free_reply_state(struct ptlrpc_reply_state *rs)
562 PTLRPC_RS_DEBUG_LRU_DEL(rs);
564 LASSERT(atomic_read(&rs->rs_refcount) == 0);
565 LASSERT(!rs->rs_difficult || rs->rs_handled);
566 LASSERT(!rs->rs_on_net);
567 LASSERT(!rs->rs_scheduled);
568 LASSERT(rs->rs_export == NULL);
569 LASSERT(rs->rs_nlocks == 0);
570 LASSERT(list_empty(&rs->rs_exp_list));
571 LASSERT(list_empty(&rs->rs_obd_list));
573 sptlrpc_svc_free_rs(rs);
576 static int lustre_unpack_msg_v2(struct lustre_msg_v2 *m, int len)
578 int swabbed, required_len, i, buflen;
580 /* Now we know the sender speaks my language. */
581 required_len = lustre_msg_hdr_size_v2(0);
582 if (len < required_len) {
583 /* can't even look inside the message */
584 CERROR("message length %d too small for lustre_msg\n", len);
588 swabbed = (m->lm_magic == LUSTRE_MSG_MAGIC_V2_SWABBED);
591 __swab32s(&m->lm_magic);
592 __swab32s(&m->lm_bufcount);
593 __swab32s(&m->lm_secflvr);
594 __swab32s(&m->lm_repsize);
595 __swab32s(&m->lm_cksum);
596 __swab32s(&m->lm_flags);
597 BUILD_BUG_ON(offsetof(typeof(*m), lm_padding_2) == 0);
598 BUILD_BUG_ON(offsetof(typeof(*m), lm_padding_3) == 0);
601 if (m->lm_bufcount == 0 || m->lm_bufcount > PTLRPC_MAX_BUFCOUNT) {
602 CERROR("message bufcount %d is not valid\n", m->lm_bufcount);
605 required_len = lustre_msg_hdr_size_v2(m->lm_bufcount);
606 if (len < required_len) {
607 /* didn't receive all the buffer lengths */
608 CERROR("message length %d too small for %d buflens\n",
609 len, m->lm_bufcount);
613 for (i = 0; i < m->lm_bufcount; i++) {
615 __swab32s(&m->lm_buflens[i]);
616 buflen = cfs_size_round(m->lm_buflens[i]);
617 if (buflen < 0 || buflen > PTLRPC_MAX_BUFLEN) {
618 CERROR("buffer %d length %d is not valid\n", i, buflen);
621 required_len += buflen;
623 if (len < required_len || required_len > PTLRPC_MAX_BUFLEN) {
624 CERROR("len: %d, required_len %d, bufcount: %d\n",
625 len, required_len, m->lm_bufcount);
626 for (i = 0; i < m->lm_bufcount; i++)
627 CERROR("buffer %d length %d\n", i, m->lm_buflens[i]);
634 int __lustre_unpack_msg(struct lustre_msg *m, int len)
636 int required_len, rc;
640 * We can provide a slightly better error log, if we check the
641 * message magic and version first. In the future, struct
642 * lustre_msg may grow, and we'd like to log a version mismatch,
643 * rather than a short message.
645 required_len = offsetof(struct lustre_msg, lm_magic) +
647 if (len < required_len) {
648 /* can't even look inside the message */
649 CERROR("message length %d too small for magic/version check\n",
654 rc = lustre_unpack_msg_v2(m, len);
658 EXPORT_SYMBOL(__lustre_unpack_msg);
660 int ptlrpc_unpack_req_msg(struct ptlrpc_request *req, int len)
664 rc = __lustre_unpack_msg(req->rq_reqmsg, len);
666 lustre_set_req_swabbed(req, MSG_PTLRPC_HEADER_OFF);
672 int ptlrpc_unpack_rep_msg(struct ptlrpc_request *req, int len)
676 rc = __lustre_unpack_msg(req->rq_repmsg, len);
678 lustre_set_rep_swabbed(req, MSG_PTLRPC_HEADER_OFF);
684 static inline int lustre_unpack_ptlrpc_body_v2(struct ptlrpc_request *req,
685 const int inout, int offset)
687 struct ptlrpc_body *pb;
688 struct lustre_msg_v2 *m = inout ? req->rq_reqmsg : req->rq_repmsg;
690 pb = lustre_msg_buf_v2(m, offset, sizeof(struct ptlrpc_body_v2));
692 CERROR("error unpacking ptlrpc body\n");
695 if (ptlrpc_buf_need_swab(req, inout, offset)) {
696 lustre_swab_ptlrpc_body(pb);
697 ptlrpc_buf_set_swabbed(req, inout, offset);
700 if ((pb->pb_version & ~LUSTRE_VERSION_MASK) != PTLRPC_MSG_VERSION) {
701 CERROR("wrong lustre_msg version %08x\n", pb->pb_version);
706 pb->pb_status = ptlrpc_status_ntoh(pb->pb_status);
711 int lustre_unpack_req_ptlrpc_body(struct ptlrpc_request *req, int offset)
713 switch (req->rq_reqmsg->lm_magic) {
714 case LUSTRE_MSG_MAGIC_V2:
715 return lustre_unpack_ptlrpc_body_v2(req, 1, offset);
717 CERROR("bad lustre msg magic: %08x\n",
718 req->rq_reqmsg->lm_magic);
723 int lustre_unpack_rep_ptlrpc_body(struct ptlrpc_request *req, int offset)
725 switch (req->rq_repmsg->lm_magic) {
726 case LUSTRE_MSG_MAGIC_V2:
727 return lustre_unpack_ptlrpc_body_v2(req, 0, offset);
729 CERROR("bad lustre msg magic: %08x\n",
730 req->rq_repmsg->lm_magic);
735 static inline __u32 lustre_msg_buflen_v2(struct lustre_msg_v2 *m, __u32 n)
737 if (n >= m->lm_bufcount)
740 return m->lm_buflens[n];
744 * lustre_msg_buflen - return the length of buffer \a n in message \a m
745 * \param m lustre_msg (request or reply) to look at
746 * \param n message index (base 0)
748 * returns zero for non-existent message indices
750 __u32 lustre_msg_buflen(struct lustre_msg *m, __u32 n)
752 switch (m->lm_magic) {
753 case LUSTRE_MSG_MAGIC_V2:
754 return lustre_msg_buflen_v2(m, n);
756 CERROR("incorrect message magic: %08x\n", m->lm_magic);
760 EXPORT_SYMBOL(lustre_msg_buflen);
763 lustre_msg_set_buflen_v2(struct lustre_msg_v2 *m, __u32 n, __u32 len)
765 if (n >= m->lm_bufcount)
768 m->lm_buflens[n] = len;
771 void lustre_msg_set_buflen(struct lustre_msg *m, __u32 n, __u32 len)
773 switch (m->lm_magic) {
774 case LUSTRE_MSG_MAGIC_V2:
775 lustre_msg_set_buflen_v2(m, n, len);
778 LASSERTF(0, "incorrect message magic: %08x\n", m->lm_magic);
783 * NB return the bufcount for lustre_msg_v2 format, so if message is packed
784 * in V1 format, the result is one bigger. (add struct ptlrpc_body).
786 __u32 lustre_msg_bufcount(struct lustre_msg *m)
788 switch (m->lm_magic) {
789 case LUSTRE_MSG_MAGIC_V2:
790 return m->lm_bufcount;
792 CERROR("incorrect message magic: %08x\n", m->lm_magic);
797 char *lustre_msg_string(struct lustre_msg *m, __u32 index, __u32 max_len)
799 /* max_len == 0 means the string should fill the buffer */
803 switch (m->lm_magic) {
804 case LUSTRE_MSG_MAGIC_V2:
805 str = lustre_msg_buf_v2(m, index, 0);
806 blen = lustre_msg_buflen_v2(m, index);
809 LASSERTF(0, "incorrect message magic: %08x\n", m->lm_magic);
813 CERROR("can't unpack string in msg %p buffer[%d]\n", m, index);
817 slen = strnlen(str, blen);
819 if (slen == blen) { /* not NULL terminated */
820 CERROR("can't unpack non-NULL terminated string in msg %p buffer[%d] len %d\n",
824 if (blen > PTLRPC_MAX_BUFLEN) {
825 CERROR("buffer length of msg %p buffer[%d] is invalid(%d)\n",
831 if (slen != blen - 1) {
832 CERROR("can't unpack short string in msg %p buffer[%d] len %d: strlen %d\n",
833 m, index, blen, slen);
836 } else if (slen > max_len) {
837 CERROR("can't unpack oversized string in msg %p buffer[%d] len %d strlen %d: max %d expected\n",
838 m, index, blen, slen, max_len);
845 /* Wrap up the normal fixed length cases */
846 static inline void *__lustre_swab_buf(struct lustre_msg *msg, __u32 index,
847 __u32 min_size, void *swabber)
851 LASSERT(msg != NULL);
852 switch (msg->lm_magic) {
853 case LUSTRE_MSG_MAGIC_V2:
854 ptr = lustre_msg_buf_v2(msg, index, min_size);
857 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
860 if (ptr != NULL && swabber != NULL)
861 ((void (*)(void *))swabber)(ptr);
866 static inline struct ptlrpc_body *lustre_msg_ptlrpc_body(struct lustre_msg *msg)
868 return lustre_msg_buf_v2(msg, MSG_PTLRPC_BODY_OFF,
869 sizeof(struct ptlrpc_body_v2));
872 enum lustre_msghdr lustre_msghdr_get_flags(struct lustre_msg *msg)
874 switch (msg->lm_magic) {
875 case LUSTRE_MSG_MAGIC_V2:
876 /* already in host endian */
877 return msg->lm_flags;
879 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
883 EXPORT_SYMBOL(lustre_msghdr_get_flags);
885 void lustre_msghdr_set_flags(struct lustre_msg *msg, __u32 flags)
887 switch (msg->lm_magic) {
888 case LUSTRE_MSG_MAGIC_V2:
889 msg->lm_flags = flags;
892 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
896 __u32 lustre_msg_get_flags(struct lustre_msg *msg)
898 switch (msg->lm_magic) {
899 case LUSTRE_MSG_MAGIC_V2: {
900 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
904 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
909 * flags might be printed in debug code while message
915 EXPORT_SYMBOL(lustre_msg_get_flags);
917 void lustre_msg_add_flags(struct lustre_msg *msg, __u32 flags)
919 switch (msg->lm_magic) {
920 case LUSTRE_MSG_MAGIC_V2: {
921 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
922 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
923 pb->pb_flags |= flags;
927 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
930 EXPORT_SYMBOL(lustre_msg_add_flags);
932 void lustre_msg_set_flags(struct lustre_msg *msg, __u32 flags)
934 switch (msg->lm_magic) {
935 case LUSTRE_MSG_MAGIC_V2: {
936 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
937 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
938 pb->pb_flags = flags;
942 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
946 void lustre_msg_clear_flags(struct lustre_msg *msg, __u32 flags)
948 switch (msg->lm_magic) {
949 case LUSTRE_MSG_MAGIC_V2: {
950 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
951 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
952 pb->pb_flags &= ~flags;
957 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
960 EXPORT_SYMBOL(lustre_msg_clear_flags);
962 __u32 lustre_msg_get_op_flags(struct lustre_msg *msg)
964 switch (msg->lm_magic) {
965 case LUSTRE_MSG_MAGIC_V2: {
966 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
968 return pb->pb_op_flags;
970 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
978 void lustre_msg_add_op_flags(struct lustre_msg *msg, __u32 flags)
980 switch (msg->lm_magic) {
981 case LUSTRE_MSG_MAGIC_V2: {
982 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
983 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
984 pb->pb_op_flags |= flags;
988 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
991 EXPORT_SYMBOL(lustre_msg_add_op_flags);
993 struct lustre_handle *lustre_msg_get_handle(struct lustre_msg *msg)
995 switch (msg->lm_magic) {
996 case LUSTRE_MSG_MAGIC_V2: {
997 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
999 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1002 return &pb->pb_handle;
1005 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1010 __u32 lustre_msg_get_type(struct lustre_msg *msg)
1012 switch (msg->lm_magic) {
1013 case LUSTRE_MSG_MAGIC_V2: {
1014 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1016 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1017 return PTL_RPC_MSG_ERR;
1022 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1023 return PTL_RPC_MSG_ERR;
1026 EXPORT_SYMBOL(lustre_msg_get_type);
1028 enum lustre_msg_version lustre_msg_get_version(struct lustre_msg *msg)
1030 switch (msg->lm_magic) {
1031 case LUSTRE_MSG_MAGIC_V2: {
1032 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1034 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1037 return pb->pb_version;
1040 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1045 void lustre_msg_add_version(struct lustre_msg *msg, __u32 version)
1047 switch (msg->lm_magic) {
1048 case LUSTRE_MSG_MAGIC_V2: {
1049 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1050 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1051 pb->pb_version |= version;
1055 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1059 __u32 lustre_msg_get_opc(struct lustre_msg *msg)
1061 switch (msg->lm_magic) {
1062 case LUSTRE_MSG_MAGIC_V2: {
1063 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1065 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1071 CERROR("incorrect message magic: %08x (msg:%p)\n",
1072 msg->lm_magic, msg);
1076 EXPORT_SYMBOL(lustre_msg_get_opc);
1078 __u64 lustre_msg_get_last_xid(struct lustre_msg *msg)
1080 switch (msg->lm_magic) {
1081 case LUSTRE_MSG_MAGIC_V2: {
1082 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1084 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1087 return pb->pb_last_xid;
1090 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1094 EXPORT_SYMBOL(lustre_msg_get_last_xid);
1096 __u16 lustre_msg_get_tag(struct lustre_msg *msg)
1098 switch (msg->lm_magic) {
1099 case LUSTRE_MSG_MAGIC_V2: {
1100 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1102 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1108 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1112 EXPORT_SYMBOL(lustre_msg_get_tag);
1114 __u64 lustre_msg_get_last_committed(struct lustre_msg *msg)
1116 switch (msg->lm_magic) {
1117 case LUSTRE_MSG_MAGIC_V2: {
1118 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1120 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1123 return pb->pb_last_committed;
1126 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1130 EXPORT_SYMBOL(lustre_msg_get_last_committed);
1132 __u64 *lustre_msg_get_versions(struct lustre_msg *msg)
1134 switch (msg->lm_magic) {
1135 case LUSTRE_MSG_MAGIC_V2: {
1136 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1138 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1141 return pb->pb_pre_versions;
1144 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1148 EXPORT_SYMBOL(lustre_msg_get_versions);
1150 __u64 lustre_msg_get_transno(struct lustre_msg *msg)
1152 switch (msg->lm_magic) {
1153 case LUSTRE_MSG_MAGIC_V2: {
1154 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1156 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1159 return pb->pb_transno;
1162 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1166 EXPORT_SYMBOL(lustre_msg_get_transno);
1168 int lustre_msg_get_status(struct lustre_msg *msg)
1170 switch (msg->lm_magic) {
1171 case LUSTRE_MSG_MAGIC_V2: {
1172 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1174 return pb->pb_status;
1175 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1180 * status might be printed in debug code while message
1186 EXPORT_SYMBOL(lustre_msg_get_status);
1188 __u64 lustre_msg_get_slv(struct lustre_msg *msg)
1190 switch (msg->lm_magic) {
1191 case LUSTRE_MSG_MAGIC_V2: {
1192 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1194 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1200 CERROR("invalid msg magic %08x\n", msg->lm_magic);
1206 void lustre_msg_set_slv(struct lustre_msg *msg, __u64 slv)
1208 switch (msg->lm_magic) {
1209 case LUSTRE_MSG_MAGIC_V2: {
1210 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1212 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1219 CERROR("invalid msg magic %x\n", msg->lm_magic);
1224 __u32 lustre_msg_get_limit(struct lustre_msg *msg)
1226 switch (msg->lm_magic) {
1227 case LUSTRE_MSG_MAGIC_V2: {
1228 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1230 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1233 return pb->pb_limit;
1236 CERROR("invalid msg magic %x\n", msg->lm_magic);
1242 void lustre_msg_set_limit(struct lustre_msg *msg, __u64 limit)
1244 switch (msg->lm_magic) {
1245 case LUSTRE_MSG_MAGIC_V2: {
1246 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1248 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1251 pb->pb_limit = limit;
1255 CERROR("invalid msg magic %08x\n", msg->lm_magic);
1260 __u32 lustre_msg_get_conn_cnt(struct lustre_msg *msg)
1262 switch (msg->lm_magic) {
1263 case LUSTRE_MSG_MAGIC_V2: {
1264 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1266 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1269 return pb->pb_conn_cnt;
1272 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1276 EXPORT_SYMBOL(lustre_msg_get_conn_cnt);
1278 __u32 lustre_msg_get_magic(struct lustre_msg *msg)
1280 switch (msg->lm_magic) {
1281 case LUSTRE_MSG_MAGIC_V2:
1282 return msg->lm_magic;
1284 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1289 __u32 lustre_msg_get_timeout(struct lustre_msg *msg)
1291 switch (msg->lm_magic) {
1292 case LUSTRE_MSG_MAGIC_V2: {
1293 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1295 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1298 return pb->pb_timeout;
1301 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1306 __u32 lustre_msg_get_service_time(struct lustre_msg *msg)
1308 switch (msg->lm_magic) {
1309 case LUSTRE_MSG_MAGIC_V2: {
1310 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1312 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1315 return pb->pb_service_time;
1318 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1323 char *lustre_msg_get_jobid(struct lustre_msg *msg)
1325 switch (msg->lm_magic) {
1326 case LUSTRE_MSG_MAGIC_V2: {
1327 struct ptlrpc_body *pb;
1329 /* the old pltrpc_body_v2 is smaller; doesn't include jobid */
1330 if (msg->lm_buflens[MSG_PTLRPC_BODY_OFF] <
1331 sizeof(struct ptlrpc_body))
1334 pb = lustre_msg_buf_v2(msg, MSG_PTLRPC_BODY_OFF,
1335 sizeof(struct ptlrpc_body));
1339 return pb->pb_jobid;
1342 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1346 EXPORT_SYMBOL(lustre_msg_get_jobid);
1348 __u32 lustre_msg_get_cksum(struct lustre_msg *msg)
1350 switch (msg->lm_magic) {
1351 case LUSTRE_MSG_MAGIC_V2:
1352 return msg->lm_cksum;
1354 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1359 __u64 lustre_msg_get_mbits(struct lustre_msg *msg)
1361 switch (msg->lm_magic) {
1362 case LUSTRE_MSG_MAGIC_V2: {
1363 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1365 CERROR("invalid msg %p: no ptlrpc body!\n", msg);
1368 return pb->pb_mbits;
1371 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1376 __u32 lustre_msg_calc_cksum(struct lustre_msg *msg)
1378 switch (msg->lm_magic) {
1379 case LUSTRE_MSG_MAGIC_V2: {
1380 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1381 __u32 len = lustre_msg_buflen(msg, MSG_PTLRPC_BODY_OFF);
1383 unsigned int hsize = 4;
1386 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1387 cfs_crypto_hash_digest(CFS_HASH_ALG_CRC32, (unsigned char *)pb,
1388 len, NULL, 0, (unsigned char *)&crc,
1393 CERROR("incorrect message magic: %08x\n", msg->lm_magic);
1398 void lustre_msg_set_handle(struct lustre_msg *msg, struct lustre_handle *handle)
1400 switch (msg->lm_magic) {
1401 case LUSTRE_MSG_MAGIC_V2: {
1402 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1403 LASSERTF(pb, "invalid msg %p: no ptlrpc body!\n", msg);
1404 pb->pb_handle = *handle;
1408 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1412 void lustre_msg_set_type(struct lustre_msg *msg, __u32 type)
1414 switch (msg->lm_magic) {
1415 case LUSTRE_MSG_MAGIC_V2: {
1416 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1417 LASSERTF(pb, "invalid msg %p: no ptlrpc body!\n", msg);
1422 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1426 void lustre_msg_set_opc(struct lustre_msg *msg, __u32 opc)
1428 switch (msg->lm_magic) {
1429 case LUSTRE_MSG_MAGIC_V2: {
1430 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1431 LASSERTF(pb, "invalid msg %p: no ptlrpc body!\n", msg);
1436 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1440 void lustre_msg_set_last_xid(struct lustre_msg *msg, __u64 last_xid)
1442 switch (msg->lm_magic) {
1443 case LUSTRE_MSG_MAGIC_V2: {
1444 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1445 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1446 pb->pb_last_xid = last_xid;
1450 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1453 EXPORT_SYMBOL(lustre_msg_set_last_xid);
1455 void lustre_msg_set_tag(struct lustre_msg *msg, __u16 tag)
1457 switch (msg->lm_magic) {
1458 case LUSTRE_MSG_MAGIC_V2: {
1459 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1460 LASSERTF(pb, "invalid msg %p: no ptlrpc body!\n", msg);
1465 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1468 EXPORT_SYMBOL(lustre_msg_set_tag);
1470 void lustre_msg_set_last_committed(struct lustre_msg *msg, __u64 last_committed)
1472 switch (msg->lm_magic) {
1473 case LUSTRE_MSG_MAGIC_V2: {
1474 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1475 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1476 pb->pb_last_committed = last_committed;
1480 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1484 void lustre_msg_set_versions(struct lustre_msg *msg, __u64 *versions)
1486 switch (msg->lm_magic) {
1487 case LUSTRE_MSG_MAGIC_V2: {
1488 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1489 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1490 pb->pb_pre_versions[0] = versions[0];
1491 pb->pb_pre_versions[1] = versions[1];
1492 pb->pb_pre_versions[2] = versions[2];
1493 pb->pb_pre_versions[3] = versions[3];
1497 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1500 EXPORT_SYMBOL(lustre_msg_set_versions);
1502 void lustre_msg_set_transno(struct lustre_msg *msg, __u64 transno)
1504 switch (msg->lm_magic) {
1505 case LUSTRE_MSG_MAGIC_V2: {
1506 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1507 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1508 pb->pb_transno = transno;
1512 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1515 EXPORT_SYMBOL(lustre_msg_set_transno);
1517 void lustre_msg_set_status(struct lustre_msg *msg, __u32 status)
1519 switch (msg->lm_magic) {
1520 case LUSTRE_MSG_MAGIC_V2: {
1521 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1522 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1523 pb->pb_status = status;
1527 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1530 EXPORT_SYMBOL(lustre_msg_set_status);
1532 void lustre_msg_set_conn_cnt(struct lustre_msg *msg, __u32 conn_cnt)
1534 switch (msg->lm_magic) {
1535 case LUSTRE_MSG_MAGIC_V2: {
1536 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1537 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1538 pb->pb_conn_cnt = conn_cnt;
1542 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1546 void lustre_msg_set_timeout(struct lustre_msg *msg, __u32 timeout)
1548 switch (msg->lm_magic) {
1549 case LUSTRE_MSG_MAGIC_V2: {
1550 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1551 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1552 pb->pb_timeout = timeout;
1556 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1560 void lustre_msg_set_service_time(struct lustre_msg *msg, __u32 service_time)
1562 switch (msg->lm_magic) {
1563 case LUSTRE_MSG_MAGIC_V2: {
1564 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1565 LASSERTF(pb, "invalid msg %p: no ptlrpc body!\n", msg);
1566 pb->pb_service_time = service_time;
1570 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1574 void lustre_msg_set_jobid(struct lustre_msg *msg, char *jobid)
1576 switch (msg->lm_magic) {
1577 case LUSTRE_MSG_MAGIC_V2: {
1578 __u32 opc = lustre_msg_get_opc(msg);
1579 struct ptlrpc_body *pb;
1581 /* Don't set jobid for ldlm ast RPCs, they've been shrinked.
1582 * See the comment in ptlrpc_request_pack(). */
1583 if (!opc || opc == LDLM_BL_CALLBACK ||
1584 opc == LDLM_CP_CALLBACK || opc == LDLM_GL_CALLBACK)
1587 pb = lustre_msg_buf_v2(msg, MSG_PTLRPC_BODY_OFF,
1588 sizeof(struct ptlrpc_body));
1589 LASSERTF(pb, "invalid msg %p: no ptlrpc body!\n", msg);
1592 memcpy(pb->pb_jobid, jobid, sizeof(pb->pb_jobid));
1593 else if (pb->pb_jobid[0] == '\0')
1594 lustre_get_jobid(pb->pb_jobid, sizeof(pb->pb_jobid));
1598 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1601 EXPORT_SYMBOL(lustre_msg_set_jobid);
1603 void lustre_msg_set_cksum(struct lustre_msg *msg, __u32 cksum)
1605 switch (msg->lm_magic) {
1606 case LUSTRE_MSG_MAGIC_V2:
1607 msg->lm_cksum = cksum;
1610 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1614 void lustre_msg_set_mbits(struct lustre_msg *msg, __u64 mbits)
1616 switch (msg->lm_magic) {
1617 case LUSTRE_MSG_MAGIC_V2: {
1618 struct ptlrpc_body *pb = lustre_msg_ptlrpc_body(msg);
1620 LASSERTF(pb != NULL, "invalid msg %p: no ptlrpc body!\n", msg);
1621 pb->pb_mbits = mbits;
1625 LASSERTF(0, "incorrect message magic: %08x\n", msg->lm_magic);
1629 void ptlrpc_request_set_replen(struct ptlrpc_request *req)
1631 int count = req_capsule_filled_sizes(&req->rq_pill, RCL_SERVER);
1633 req->rq_replen = lustre_msg_size(req->rq_reqmsg->lm_magic, count,
1634 req->rq_pill.rc_area[RCL_SERVER]);
1635 if (req->rq_reqmsg->lm_magic == LUSTRE_MSG_MAGIC_V2)
1636 req->rq_reqmsg->lm_repsize = req->rq_replen;
1638 EXPORT_SYMBOL(ptlrpc_request_set_replen);
1640 void ptlrpc_req_set_repsize(struct ptlrpc_request *req, int count, __u32 *lens)
1642 req->rq_replen = lustre_msg_size(req->rq_reqmsg->lm_magic, count, lens);
1643 if (req->rq_reqmsg->lm_magic == LUSTRE_MSG_MAGIC_V2)
1644 req->rq_reqmsg->lm_repsize = req->rq_replen;
1648 * Send a remote set_info_async.
1650 * This may go from client to server or server to client.
1652 int do_set_info_async(struct obd_import *imp,
1653 int opcode, int version,
1654 size_t keylen, void *key,
1655 size_t vallen, void *val,
1656 struct ptlrpc_request_set *set)
1658 struct ptlrpc_request *req;
1663 req = ptlrpc_request_alloc(imp, &RQF_OBD_SET_INFO);
1667 req_capsule_set_size(&req->rq_pill, &RMF_SETINFO_KEY,
1668 RCL_CLIENT, keylen);
1669 req_capsule_set_size(&req->rq_pill, &RMF_SETINFO_VAL,
1670 RCL_CLIENT, vallen);
1671 rc = ptlrpc_request_pack(req, version, opcode);
1673 ptlrpc_request_free(req);
1677 tmp = req_capsule_client_get(&req->rq_pill, &RMF_SETINFO_KEY);
1678 memcpy(tmp, key, keylen);
1679 tmp = req_capsule_client_get(&req->rq_pill, &RMF_SETINFO_VAL);
1680 memcpy(tmp, val, vallen);
1682 ptlrpc_request_set_replen(req);
1685 ptlrpc_set_add_req(set, req);
1686 ptlrpc_check_set(NULL, set);
1688 rc = ptlrpc_queue_wait(req);
1689 ptlrpc_req_finished(req);
1694 EXPORT_SYMBOL(do_set_info_async);
1696 /* byte flipping routines for all wire types declared in
1697 * lustre_idl.h implemented here.
1699 void lustre_swab_ptlrpc_body(struct ptlrpc_body *body)
1701 __swab32s(&body->pb_type);
1702 __swab32s(&body->pb_version);
1703 __swab32s(&body->pb_opc);
1704 __swab32s(&body->pb_status);
1705 __swab64s(&body->pb_last_xid);
1706 __swab16s(&body->pb_tag);
1707 BUILD_BUG_ON(offsetof(typeof(*body), pb_padding0) == 0);
1708 BUILD_BUG_ON(offsetof(typeof(*body), pb_padding1) == 0);
1709 __swab64s(&body->pb_last_committed);
1710 __swab64s(&body->pb_transno);
1711 __swab32s(&body->pb_flags);
1712 __swab32s(&body->pb_op_flags);
1713 __swab32s(&body->pb_conn_cnt);
1714 __swab32s(&body->pb_timeout);
1715 __swab32s(&body->pb_service_time);
1716 __swab32s(&body->pb_limit);
1717 __swab64s(&body->pb_slv);
1718 __swab64s(&body->pb_pre_versions[0]);
1719 __swab64s(&body->pb_pre_versions[1]);
1720 __swab64s(&body->pb_pre_versions[2]);
1721 __swab64s(&body->pb_pre_versions[3]);
1722 __swab64s(&body->pb_mbits);
1723 BUILD_BUG_ON(offsetof(typeof(*body), pb_padding64_0) == 0);
1724 BUILD_BUG_ON(offsetof(typeof(*body), pb_padding64_1) == 0);
1725 BUILD_BUG_ON(offsetof(typeof(*body), pb_padding64_2) == 0);
1727 * While we need to maintain compatibility between
1728 * clients and servers without ptlrpc_body_v2 (< 2.3)
1729 * do not swab any fields beyond pb_jobid, as we are
1730 * using this swab function for both ptlrpc_body
1731 * and ptlrpc_body_v2.
1733 /* pb_jobid is an ASCII string and should not be swabbed */
1734 BUILD_BUG_ON(offsetof(typeof(*body), pb_jobid) == 0);
1737 void lustre_swab_connect(struct obd_connect_data *ocd)
1739 __swab64s(&ocd->ocd_connect_flags);
1740 __swab32s(&ocd->ocd_version);
1741 __swab32s(&ocd->ocd_grant);
1742 __swab64s(&ocd->ocd_ibits_known);
1743 __swab32s(&ocd->ocd_index);
1744 __swab32s(&ocd->ocd_brw_size);
1746 * ocd_blocksize and ocd_inodespace don't need to be swabbed because
1747 * they are 8-byte values
1749 __swab16s(&ocd->ocd_grant_tax_kb);
1750 __swab32s(&ocd->ocd_grant_max_blks);
1751 __swab64s(&ocd->ocd_transno);
1752 __swab32s(&ocd->ocd_group);
1753 __swab32s(&ocd->ocd_cksum_types);
1754 __swab32s(&ocd->ocd_instance);
1756 * Fields after ocd_cksum_types are only accessible by the receiver
1757 * if the corresponding flag in ocd_connect_flags is set. Accessing
1758 * any field after ocd_maxbytes on the receiver without a valid flag
1759 * may result in out-of-bound memory access and kernel oops.
1761 if (ocd->ocd_connect_flags & OBD_CONNECT_MAX_EASIZE)
1762 __swab32s(&ocd->ocd_max_easize);
1763 if (ocd->ocd_connect_flags & OBD_CONNECT_MAXBYTES)
1764 __swab64s(&ocd->ocd_maxbytes);
1765 if (ocd->ocd_connect_flags & OBD_CONNECT_MULTIMODRPCS)
1766 __swab16s(&ocd->ocd_maxmodrpcs);
1767 BUILD_BUG_ON(offsetof(typeof(*ocd), padding0) == 0);
1768 BUILD_BUG_ON(offsetof(typeof(*ocd), padding1) == 0);
1769 if (ocd->ocd_connect_flags & OBD_CONNECT_FLAGS2)
1770 __swab64s(&ocd->ocd_connect_flags2);
1771 BUILD_BUG_ON(offsetof(typeof(*ocd), padding3) == 0);
1772 BUILD_BUG_ON(offsetof(typeof(*ocd), padding4) == 0);
1773 BUILD_BUG_ON(offsetof(typeof(*ocd), padding5) == 0);
1774 BUILD_BUG_ON(offsetof(typeof(*ocd), padding6) == 0);
1775 BUILD_BUG_ON(offsetof(typeof(*ocd), padding7) == 0);
1776 BUILD_BUG_ON(offsetof(typeof(*ocd), padding8) == 0);
1777 BUILD_BUG_ON(offsetof(typeof(*ocd), padding9) == 0);
1778 BUILD_BUG_ON(offsetof(typeof(*ocd), paddingA) == 0);
1779 BUILD_BUG_ON(offsetof(typeof(*ocd), paddingB) == 0);
1780 BUILD_BUG_ON(offsetof(typeof(*ocd), paddingC) == 0);
1781 BUILD_BUG_ON(offsetof(typeof(*ocd), paddingD) == 0);
1782 BUILD_BUG_ON(offsetof(typeof(*ocd), paddingE) == 0);
1783 BUILD_BUG_ON(offsetof(typeof(*ocd), paddingF) == 0);
1786 static void lustre_swab_ost_layout(struct ost_layout *ol)
1788 __swab32s(&ol->ol_stripe_size);
1789 __swab32s(&ol->ol_stripe_count);
1790 __swab64s(&ol->ol_comp_start);
1791 __swab64s(&ol->ol_comp_end);
1792 __swab32s(&ol->ol_comp_id);
1795 void lustre_swab_obdo(struct obdo *o)
1797 __swab64s(&o->o_valid);
1798 lustre_swab_ost_id(&o->o_oi);
1799 __swab64s(&o->o_parent_seq);
1800 __swab64s(&o->o_size);
1801 __swab64s(&o->o_mtime);
1802 __swab64s(&o->o_atime);
1803 __swab64s(&o->o_ctime);
1804 __swab64s(&o->o_blocks);
1805 __swab64s(&o->o_grant);
1806 __swab32s(&o->o_blksize);
1807 __swab32s(&o->o_mode);
1808 __swab32s(&o->o_uid);
1809 __swab32s(&o->o_gid);
1810 __swab32s(&o->o_flags);
1811 __swab32s(&o->o_nlink);
1812 __swab32s(&o->o_parent_oid);
1813 __swab32s(&o->o_misc);
1814 __swab64s(&o->o_ioepoch);
1815 __swab32s(&o->o_stripe_idx);
1816 __swab32s(&o->o_parent_ver);
1817 lustre_swab_ost_layout(&o->o_layout);
1818 __swab32s(&o->o_layout_version);
1819 __swab32s(&o->o_uid_h);
1820 __swab32s(&o->o_gid_h);
1821 __swab64s(&o->o_data_version);
1822 __swab32s(&o->o_projid);
1823 BUILD_BUG_ON(offsetof(typeof(*o), o_padding_4) == 0);
1824 BUILD_BUG_ON(offsetof(typeof(*o), o_padding_5) == 0);
1825 BUILD_BUG_ON(offsetof(typeof(*o), o_padding_6) == 0);
1828 EXPORT_SYMBOL(lustre_swab_obdo);
1830 void lustre_swab_obd_statfs(struct obd_statfs *os)
1832 __swab64s(&os->os_type);
1833 __swab64s(&os->os_blocks);
1834 __swab64s(&os->os_bfree);
1835 __swab64s(&os->os_bavail);
1836 __swab64s(&os->os_files);
1837 __swab64s(&os->os_ffree);
1838 /* no need to swab os_fsid */
1839 __swab32s(&os->os_bsize);
1840 __swab32s(&os->os_namelen);
1841 __swab64s(&os->os_maxbytes);
1842 __swab32s(&os->os_state);
1843 __swab32s(&os->os_fprecreated);
1844 __swab32s(&os->os_granted);
1845 BUILD_BUG_ON(offsetof(typeof(*os), os_spare3) == 0);
1846 BUILD_BUG_ON(offsetof(typeof(*os), os_spare4) == 0);
1847 BUILD_BUG_ON(offsetof(typeof(*os), os_spare5) == 0);
1848 BUILD_BUG_ON(offsetof(typeof(*os), os_spare6) == 0);
1849 BUILD_BUG_ON(offsetof(typeof(*os), os_spare7) == 0);
1850 BUILD_BUG_ON(offsetof(typeof(*os), os_spare8) == 0);
1851 BUILD_BUG_ON(offsetof(typeof(*os), os_spare9) == 0);
1854 void lustre_swab_obd_ioobj(struct obd_ioobj *ioo)
1856 lustre_swab_ost_id(&ioo->ioo_oid);
1857 __swab32s(&ioo->ioo_max_brw);
1858 __swab32s(&ioo->ioo_bufcnt);
1861 void lustre_swab_niobuf_remote(struct niobuf_remote *nbr)
1863 __swab64s(&nbr->rnb_offset);
1864 __swab32s(&nbr->rnb_len);
1865 __swab32s(&nbr->rnb_flags);
1868 void lustre_swab_ost_body(struct ost_body *b)
1870 lustre_swab_obdo(&b->oa);
1873 void lustre_swab_ost_last_id(u64 *id)
1878 void lustre_swab_generic_32s(__u32 *val)
1883 void lustre_swab_gl_lquota_desc(struct ldlm_gl_lquota_desc *desc)
1885 lustre_swab_lu_fid(&desc->gl_id.qid_fid);
1886 __swab64s(&desc->gl_flags);
1887 __swab64s(&desc->gl_ver);
1888 __swab64s(&desc->gl_hardlimit);
1889 __swab64s(&desc->gl_softlimit);
1890 __swab64s(&desc->gl_time);
1891 BUILD_BUG_ON(offsetof(typeof(*desc), gl_pad2) == 0);
1893 EXPORT_SYMBOL(lustre_swab_gl_lquota_desc);
1895 void lustre_swab_gl_barrier_desc(struct ldlm_gl_barrier_desc *desc)
1897 __swab32s(&desc->lgbd_status);
1898 __swab32s(&desc->lgbd_timeout);
1899 BUILD_BUG_ON(offsetof(typeof(*desc), lgbd_padding) == 0);
1901 EXPORT_SYMBOL(lustre_swab_gl_barrier_desc);
1903 void lustre_swab_ost_lvb_v1(struct ost_lvb_v1 *lvb)
1905 __swab64s(&lvb->lvb_size);
1906 __swab64s(&lvb->lvb_mtime);
1907 __swab64s(&lvb->lvb_atime);
1908 __swab64s(&lvb->lvb_ctime);
1909 __swab64s(&lvb->lvb_blocks);
1911 EXPORT_SYMBOL(lustre_swab_ost_lvb_v1);
1913 void lustre_swab_ost_lvb(struct ost_lvb *lvb)
1915 __swab64s(&lvb->lvb_size);
1916 __swab64s(&lvb->lvb_mtime);
1917 __swab64s(&lvb->lvb_atime);
1918 __swab64s(&lvb->lvb_ctime);
1919 __swab64s(&lvb->lvb_blocks);
1920 __swab32s(&lvb->lvb_mtime_ns);
1921 __swab32s(&lvb->lvb_atime_ns);
1922 __swab32s(&lvb->lvb_ctime_ns);
1923 __swab32s(&lvb->lvb_padding);
1925 EXPORT_SYMBOL(lustre_swab_ost_lvb);
1927 void lustre_swab_lquota_lvb(struct lquota_lvb *lvb)
1929 __swab64s(&lvb->lvb_flags);
1930 __swab64s(&lvb->lvb_id_may_rel);
1931 __swab64s(&lvb->lvb_id_rel);
1932 __swab64s(&lvb->lvb_id_qunit);
1933 __swab64s(&lvb->lvb_pad1);
1935 EXPORT_SYMBOL(lustre_swab_lquota_lvb);
1937 void lustre_swab_barrier_lvb(struct barrier_lvb *lvb)
1939 __swab32s(&lvb->lvb_status);
1940 __swab32s(&lvb->lvb_index);
1941 BUILD_BUG_ON(offsetof(typeof(*lvb), lvb_padding) == 0);
1943 EXPORT_SYMBOL(lustre_swab_barrier_lvb);
1945 void lustre_swab_mdt_body(struct mdt_body *b)
1947 lustre_swab_lu_fid(&b->mbo_fid1);
1948 lustre_swab_lu_fid(&b->mbo_fid2);
1949 /* handle is opaque */
1950 __swab64s(&b->mbo_valid);
1951 __swab64s(&b->mbo_size);
1952 __swab64s(&b->mbo_mtime);
1953 __swab64s(&b->mbo_atime);
1954 __swab64s(&b->mbo_ctime);
1955 __swab64s(&b->mbo_blocks);
1956 __swab64s(&b->mbo_version);
1957 __swab64s(&b->mbo_t_state);
1958 __swab32s(&b->mbo_fsuid);
1959 __swab32s(&b->mbo_fsgid);
1960 __swab32s(&b->mbo_capability);
1961 __swab32s(&b->mbo_mode);
1962 __swab32s(&b->mbo_uid);
1963 __swab32s(&b->mbo_gid);
1964 __swab32s(&b->mbo_flags);
1965 __swab32s(&b->mbo_rdev);
1966 __swab32s(&b->mbo_nlink);
1967 __swab32s(&b->mbo_layout_gen);
1968 __swab32s(&b->mbo_suppgid);
1969 __swab32s(&b->mbo_eadatasize);
1970 __swab32s(&b->mbo_aclsize);
1971 __swab32s(&b->mbo_max_mdsize);
1972 BUILD_BUG_ON(offsetof(typeof(*b), mbo_unused3) == 0);
1973 __swab32s(&b->mbo_uid_h);
1974 __swab32s(&b->mbo_gid_h);
1975 __swab32s(&b->mbo_projid);
1976 __swab64s(&b->mbo_dom_size);
1977 __swab64s(&b->mbo_dom_blocks);
1978 BUILD_BUG_ON(offsetof(typeof(*b), mbo_padding_8) == 0);
1979 BUILD_BUG_ON(offsetof(typeof(*b), mbo_padding_9) == 0);
1980 BUILD_BUG_ON(offsetof(typeof(*b), mbo_padding_10) == 0);
1983 void lustre_swab_mdt_ioepoch(struct mdt_ioepoch *b)
1985 /* mio_open_handle is opaque */
1986 BUILD_BUG_ON(offsetof(typeof(*b), mio_unused1) == 0);
1987 BUILD_BUG_ON(offsetof(typeof(*b), mio_unused2) == 0);
1988 BUILD_BUG_ON(offsetof(typeof(*b), mio_padding) == 0);
1991 void lustre_swab_mgs_target_info(struct mgs_target_info *mti)
1995 __swab32s(&mti->mti_lustre_ver);
1996 __swab32s(&mti->mti_stripe_index);
1997 __swab32s(&mti->mti_config_ver);
1998 __swab32s(&mti->mti_flags);
1999 __swab32s(&mti->mti_instance);
2000 __swab32s(&mti->mti_nid_count);
2001 BUILD_BUG_ON(sizeof(lnet_nid_t) != sizeof(__u64));
2002 for (i = 0; i < MTI_NIDS_MAX; i++)
2003 __swab64s(&mti->mti_nids[i]);
2006 void lustre_swab_mgs_nidtbl_entry(struct mgs_nidtbl_entry *entry)
2010 __swab64s(&entry->mne_version);
2011 __swab32s(&entry->mne_instance);
2012 __swab32s(&entry->mne_index);
2013 __swab32s(&entry->mne_length);
2015 /* mne_nid_(count|type) must be one byte size because we're gonna
2016 * access it w/o swapping. */
2017 BUILD_BUG_ON(sizeof(entry->mne_nid_count) != sizeof(__u8));
2018 BUILD_BUG_ON(sizeof(entry->mne_nid_type) != sizeof(__u8));
2020 /* remove this assertion if ipv6 is supported. */
2021 LASSERT(entry->mne_nid_type == 0);
2022 for (i = 0; i < entry->mne_nid_count; i++) {
2023 BUILD_BUG_ON(sizeof(lnet_nid_t) != sizeof(__u64));
2024 __swab64s(&entry->u.nids[i]);
2027 EXPORT_SYMBOL(lustre_swab_mgs_nidtbl_entry);
2029 void lustre_swab_mgs_config_body(struct mgs_config_body *body)
2031 __swab64s(&body->mcb_offset);
2032 __swab32s(&body->mcb_units);
2033 __swab16s(&body->mcb_type);
2036 void lustre_swab_mgs_config_res(struct mgs_config_res *body)
2038 __swab64s(&body->mcr_offset);
2039 __swab64s(&body->mcr_size);
2042 static void lustre_swab_obd_dqinfo(struct obd_dqinfo *i)
2044 __swab64s(&i->dqi_bgrace);
2045 __swab64s(&i->dqi_igrace);
2046 __swab32s(&i->dqi_flags);
2047 __swab32s(&i->dqi_valid);
2050 static void lustre_swab_obd_dqblk(struct obd_dqblk *b)
2052 __swab64s(&b->dqb_ihardlimit);
2053 __swab64s(&b->dqb_isoftlimit);
2054 __swab64s(&b->dqb_curinodes);
2055 __swab64s(&b->dqb_bhardlimit);
2056 __swab64s(&b->dqb_bsoftlimit);
2057 __swab64s(&b->dqb_curspace);
2058 __swab64s(&b->dqb_btime);
2059 __swab64s(&b->dqb_itime);
2060 __swab32s(&b->dqb_valid);
2061 BUILD_BUG_ON(offsetof(typeof(*b), dqb_padding) == 0);
2064 void lustre_swab_obd_quotactl(struct obd_quotactl *q)
2066 __swab32s(&q->qc_cmd);
2067 __swab32s(&q->qc_type);
2068 __swab32s(&q->qc_id);
2069 __swab32s(&q->qc_stat);
2070 lustre_swab_obd_dqinfo(&q->qc_dqinfo);
2071 lustre_swab_obd_dqblk(&q->qc_dqblk);
2074 void lustre_swab_fid2path(struct getinfo_fid2path *gf)
2076 lustre_swab_lu_fid(&gf->gf_fid);
2077 __swab64s(&gf->gf_recno);
2078 __swab32s(&gf->gf_linkno);
2079 __swab32s(&gf->gf_pathlen);
2081 EXPORT_SYMBOL(lustre_swab_fid2path);
2083 static void lustre_swab_fiemap_extent(struct fiemap_extent *fm_extent)
2085 __swab64s(&fm_extent->fe_logical);
2086 __swab64s(&fm_extent->fe_physical);
2087 __swab64s(&fm_extent->fe_length);
2088 __swab32s(&fm_extent->fe_flags);
2089 __swab32s(&fm_extent->fe_device);
2092 static void lustre_swab_fiemap_hdr(struct fiemap *fiemap)
2094 __swab64s(&fiemap->fm_start);
2095 __swab64s(&fiemap->fm_length);
2096 __swab32s(&fiemap->fm_flags);
2097 __swab32s(&fiemap->fm_mapped_extents);
2098 __swab32s(&fiemap->fm_extent_count);
2099 __swab32s(&fiemap->fm_reserved);
2102 void lustre_swab_fiemap(struct fiemap *fiemap)
2106 lustre_swab_fiemap_hdr(fiemap);
2108 for (i = 0; i < fiemap->fm_mapped_extents; i++)
2109 lustre_swab_fiemap_extent(&fiemap->fm_extents[i]);
2112 void lustre_swab_fiemap_info_key(struct ll_fiemap_info_key *fiemap_info)
2114 lustre_swab_obdo(&fiemap_info->lfik_oa);
2115 lustre_swab_fiemap_hdr(&fiemap_info->lfik_fiemap);
2118 void lustre_swab_idx_info(struct idx_info *ii)
2120 __swab32s(&ii->ii_magic);
2121 __swab32s(&ii->ii_flags);
2122 __swab16s(&ii->ii_count);
2123 __swab32s(&ii->ii_attrs);
2124 lustre_swab_lu_fid(&ii->ii_fid);
2125 __swab64s(&ii->ii_version);
2126 __swab64s(&ii->ii_hash_start);
2127 __swab64s(&ii->ii_hash_end);
2128 __swab16s(&ii->ii_keysize);
2129 __swab16s(&ii->ii_recsize);
2132 void lustre_swab_lip_header(struct lu_idxpage *lip)
2135 __swab32s(&lip->lip_magic);
2136 __swab16s(&lip->lip_flags);
2137 __swab16s(&lip->lip_nr);
2139 EXPORT_SYMBOL(lustre_swab_lip_header);
2141 void lustre_swab_mdt_rec_reint (struct mdt_rec_reint *rr)
2143 __swab32s(&rr->rr_opcode);
2144 __swab32s(&rr->rr_cap);
2145 __swab32s(&rr->rr_fsuid);
2146 /* rr_fsuid_h is unused */
2147 __swab32s(&rr->rr_fsgid);
2148 /* rr_fsgid_h is unused */
2149 __swab32s(&rr->rr_suppgid1);
2150 /* rr_suppgid1_h is unused */
2151 __swab32s(&rr->rr_suppgid2);
2152 /* rr_suppgid2_h is unused */
2153 lustre_swab_lu_fid(&rr->rr_fid1);
2154 lustre_swab_lu_fid(&rr->rr_fid2);
2155 __swab64s(&rr->rr_mtime);
2156 __swab64s(&rr->rr_atime);
2157 __swab64s(&rr->rr_ctime);
2158 __swab64s(&rr->rr_size);
2159 __swab64s(&rr->rr_blocks);
2160 __swab32s(&rr->rr_bias);
2161 __swab32s(&rr->rr_mode);
2162 __swab32s(&rr->rr_flags);
2163 __swab32s(&rr->rr_flags_h);
2164 __swab32s(&rr->rr_umask);
2165 __swab16s(&rr->rr_mirror_id);
2167 BUILD_BUG_ON(offsetof(typeof(*rr), rr_padding_4) == 0);
2170 void lustre_swab_lov_desc(struct lov_desc *ld)
2172 __swab32s(&ld->ld_tgt_count);
2173 __swab32s(&ld->ld_active_tgt_count);
2174 __swab32s(&ld->ld_default_stripe_count);
2175 __swab32s(&ld->ld_pattern);
2176 __swab64s(&ld->ld_default_stripe_size);
2177 __swab64s(&ld->ld_default_stripe_offset);
2178 __swab32s(&ld->ld_qos_maxage);
2179 /* uuid endian insensitive */
2181 EXPORT_SYMBOL(lustre_swab_lov_desc);
2183 void lustre_swab_lmv_desc(struct lmv_desc *ld)
2185 __swab32s(&ld->ld_tgt_count);
2186 __swab32s(&ld->ld_active_tgt_count);
2187 __swab32s(&ld->ld_default_stripe_count);
2188 __swab32s(&ld->ld_pattern);
2189 __swab64s(&ld->ld_default_hash_size);
2190 __swab32s(&ld->ld_qos_maxage);
2191 /* uuid endian insensitive */
2194 /* This structure is always in little-endian */
2195 static void lustre_swab_lmv_mds_md_v1(struct lmv_mds_md_v1 *lmm1)
2199 __swab32s(&lmm1->lmv_magic);
2200 __swab32s(&lmm1->lmv_stripe_count);
2201 __swab32s(&lmm1->lmv_master_mdt_index);
2202 __swab32s(&lmm1->lmv_hash_type);
2203 __swab32s(&lmm1->lmv_layout_version);
2204 for (i = 0; i < lmm1->lmv_stripe_count; i++)
2205 lustre_swab_lu_fid(&lmm1->lmv_stripe_fids[i]);
2208 void lustre_swab_lmv_mds_md(union lmv_mds_md *lmm)
2210 switch (lmm->lmv_magic) {
2212 lustre_swab_lmv_mds_md_v1(&lmm->lmv_md_v1);
2218 EXPORT_SYMBOL(lustre_swab_lmv_mds_md);
2220 void lustre_swab_lmv_user_md_objects(struct lmv_user_mds_data *lmd,
2225 for (i = 0; i < stripe_count; i++)
2226 __swab32s(&(lmd[i].lum_mds));
2228 EXPORT_SYMBOL(lustre_swab_lmv_user_md_objects);
2231 void lustre_swab_lmv_user_md(struct lmv_user_md *lum)
2235 if (lum->lum_magic == LMV_MAGIC_FOREIGN) {
2236 __swab32s(&lum->lum_magic);
2237 __swab32s(&((struct lmv_foreign_md *)lum)->lfm_length);
2238 __swab32s(&((struct lmv_foreign_md *)lum)->lfm_type);
2239 __swab32s(&((struct lmv_foreign_md *)lum)->lfm_flags);
2243 count = lum->lum_stripe_count;
2244 __swab32s(&lum->lum_magic);
2245 __swab32s(&lum->lum_stripe_count);
2246 __swab32s(&lum->lum_stripe_offset);
2247 __swab32s(&lum->lum_hash_type);
2248 __swab32s(&lum->lum_type);
2249 BUILD_BUG_ON(offsetof(typeof(*lum), lum_padding1) == 0);
2250 switch (lum->lum_magic) {
2251 case LMV_USER_MAGIC_SPECIFIC:
2252 count = lum->lum_stripe_count;
2254 case __swab32(LMV_USER_MAGIC_SPECIFIC):
2255 lustre_swab_lmv_user_md_objects(lum->lum_objects, count);
2261 EXPORT_SYMBOL(lustre_swab_lmv_user_md);
2263 static void lustre_print_v1v3(unsigned int lvl, struct lov_user_md *lum,
2266 CDEBUG(lvl, "%s lov_user_md %p:\n", msg, lum);
2267 CDEBUG(lvl, "\tlmm_magic: %#x\n", lum->lmm_magic);
2268 CDEBUG(lvl, "\tlmm_pattern: %#x\n", lum->lmm_pattern);
2269 CDEBUG(lvl, "\tlmm_object_id: %llu\n", lmm_oi_id(&lum->lmm_oi));
2270 CDEBUG(lvl, "\tlmm_object_gr: %llu\n", lmm_oi_seq(&lum->lmm_oi));
2271 CDEBUG(lvl, "\tlmm_stripe_size: %#x\n", lum->lmm_stripe_size);
2272 CDEBUG(lvl, "\tlmm_stripe_count: %#x\n", lum->lmm_stripe_count);
2273 CDEBUG(lvl, "\tlmm_stripe_offset/lmm_layout_gen: %#x\n",
2274 lum->lmm_stripe_offset);
2275 if (lum->lmm_magic == LOV_USER_MAGIC_V3) {
2276 struct lov_user_md_v3 *v3 = (void *)lum;
2277 CDEBUG(lvl, "\tlmm_pool_name: %s\n", v3->lmm_pool_name);
2279 if (lum->lmm_magic == LOV_USER_MAGIC_SPECIFIC) {
2280 struct lov_user_md_v3 *v3 = (void *)lum;
2283 if (v3->lmm_pool_name[0] != '\0')
2284 CDEBUG(lvl, "\tlmm_pool_name: %s\n", v3->lmm_pool_name);
2286 CDEBUG(lvl, "\ttarget list:\n");
2287 for (i = 0; i < v3->lmm_stripe_count; i++)
2288 CDEBUG(lvl, "\t\t%u\n", v3->lmm_objects[i].l_ost_idx);
2292 void lustre_print_user_md(unsigned int lvl, struct lov_user_md *lum,
2295 struct lov_comp_md_v1 *comp_v1;
2298 if (likely(!cfs_cdebug_show(lvl, DEBUG_SUBSYSTEM)))
2301 if (lum->lmm_magic == LOV_USER_MAGIC_V1 ||
2302 lum->lmm_magic == LOV_USER_MAGIC_V3) {
2303 lustre_print_v1v3(lvl, lum, msg);
2307 if (lum->lmm_magic != LOV_USER_MAGIC_COMP_V1) {
2308 CDEBUG(lvl, "%s: bad magic: %x\n", msg, lum->lmm_magic);
2312 comp_v1 = (struct lov_comp_md_v1 *)lum;
2313 CDEBUG(lvl, "%s: lov_comp_md_v1 %p:\n", msg, lum);
2314 CDEBUG(lvl, "\tlcm_magic: %#x\n", comp_v1->lcm_magic);
2315 CDEBUG(lvl, "\tlcm_size: %#x\n", comp_v1->lcm_size);
2316 CDEBUG(lvl, "\tlcm_layout_gen: %#x\n", comp_v1->lcm_layout_gen);
2317 CDEBUG(lvl, "\tlcm_flags: %#x\n", comp_v1->lcm_flags);
2318 CDEBUG(lvl, "\tlcm_entry_count: %#x\n\n", comp_v1->lcm_entry_count);
2319 CDEBUG(lvl, "\tlcm_mirror_count: %#x\n\n", comp_v1->lcm_mirror_count);
2321 for (i = 0; i < comp_v1->lcm_entry_count; i++) {
2322 struct lov_comp_md_entry_v1 *ent = &comp_v1->lcm_entries[i];
2323 struct lov_user_md *v1;
2325 CDEBUG(lvl, "\tentry %d:\n", i);
2326 CDEBUG(lvl, "\tlcme_id: %#x\n", ent->lcme_id);
2327 CDEBUG(lvl, "\tlcme_flags: %#x\n", ent->lcme_flags);
2328 if (ent->lcme_flags & LCME_FL_NOSYNC)
2329 CDEBUG(lvl, "\tlcme_timestamp: %llu\n",
2330 ent->lcme_timestamp);
2331 CDEBUG(lvl, "\tlcme_extent.e_start: %llu\n",
2332 ent->lcme_extent.e_start);
2333 CDEBUG(lvl, "\tlcme_extent.e_end: %llu\n",
2334 ent->lcme_extent.e_end);
2335 CDEBUG(lvl, "\tlcme_offset: %#x\n", ent->lcme_offset);
2336 CDEBUG(lvl, "\tlcme_size: %#x\n\n", ent->lcme_size);
2338 v1 = (struct lov_user_md *)((char *)comp_v1 +
2339 comp_v1->lcm_entries[i].lcme_offset);
2340 lustre_print_v1v3(lvl, v1, msg);
2343 EXPORT_SYMBOL(lustre_print_user_md);
2345 static void lustre_swab_lmm_oi(struct ost_id *oi)
2347 __swab64s(&oi->oi.oi_id);
2348 __swab64s(&oi->oi.oi_seq);
2351 static void lustre_swab_lov_user_md_common(struct lov_user_md_v1 *lum)
2354 __swab32s(&lum->lmm_magic);
2355 __swab32s(&lum->lmm_pattern);
2356 lustre_swab_lmm_oi(&lum->lmm_oi);
2357 __swab32s(&lum->lmm_stripe_size);
2358 __swab16s(&lum->lmm_stripe_count);
2359 __swab16s(&lum->lmm_stripe_offset);
2363 void lustre_swab_lov_user_md_v1(struct lov_user_md_v1 *lum)
2366 CDEBUG(D_IOCTL, "swabbing lov_user_md v1\n");
2367 lustre_swab_lov_user_md_common(lum);
2370 EXPORT_SYMBOL(lustre_swab_lov_user_md_v1);
2372 void lustre_swab_lov_user_md_v3(struct lov_user_md_v3 *lum)
2375 CDEBUG(D_IOCTL, "swabbing lov_user_md v3\n");
2376 lustre_swab_lov_user_md_common((struct lov_user_md_v1 *)lum);
2377 /* lmm_pool_name nothing to do with char */
2380 EXPORT_SYMBOL(lustre_swab_lov_user_md_v3);
2382 void lustre_swab_lov_comp_md_v1(struct lov_comp_md_v1 *lum)
2384 struct lov_comp_md_entry_v1 *ent;
2385 struct lov_user_md_v1 *v1;
2386 struct lov_user_md_v3 *v3;
2390 __u16 ent_count, stripe_count;
2393 cpu_endian = lum->lcm_magic == LOV_USER_MAGIC_COMP_V1;
2394 ent_count = lum->lcm_entry_count;
2396 __swab16s(&ent_count);
2398 CDEBUG(D_IOCTL, "swabbing lov_user_comp_md v1\n");
2399 __swab32s(&lum->lcm_magic);
2400 __swab32s(&lum->lcm_size);
2401 __swab32s(&lum->lcm_layout_gen);
2402 __swab16s(&lum->lcm_flags);
2403 __swab16s(&lum->lcm_entry_count);
2404 __swab16s(&lum->lcm_mirror_count);
2405 BUILD_BUG_ON(offsetof(typeof(*lum), lcm_padding1) == 0);
2406 BUILD_BUG_ON(offsetof(typeof(*lum), lcm_padding2) == 0);
2408 for (i = 0; i < ent_count; i++) {
2409 ent = &lum->lcm_entries[i];
2410 off = ent->lcme_offset;
2411 size = ent->lcme_size;
2417 __swab32s(&ent->lcme_id);
2418 __swab32s(&ent->lcme_flags);
2419 __swab64s(&ent->lcme_timestamp);
2420 __swab64s(&ent->lcme_extent.e_start);
2421 __swab64s(&ent->lcme_extent.e_end);
2422 __swab32s(&ent->lcme_offset);
2423 __swab32s(&ent->lcme_size);
2424 __swab32s(&ent->lcme_layout_gen);
2425 BUILD_BUG_ON(offsetof(typeof(*ent), lcme_padding_1) == 0);
2427 v1 = (struct lov_user_md_v1 *)((char *)lum + off);
2428 stripe_count = v1->lmm_stripe_count;
2430 __swab16s(&stripe_count);
2432 if (v1->lmm_magic == __swab32(LOV_USER_MAGIC_V1) ||
2433 v1->lmm_magic == LOV_USER_MAGIC_V1) {
2434 lustre_swab_lov_user_md_v1(v1);
2435 if (size > sizeof(*v1))
2436 lustre_swab_lov_user_md_objects(v1->lmm_objects,
2438 } else if (v1->lmm_magic == __swab32(LOV_USER_MAGIC_V3) ||
2439 v1->lmm_magic == LOV_USER_MAGIC_V3 ||
2440 v1->lmm_magic == __swab32(LOV_USER_MAGIC_SPECIFIC) ||
2441 v1->lmm_magic == LOV_USER_MAGIC_SPECIFIC) {
2442 v3 = (struct lov_user_md_v3 *)v1;
2443 lustre_swab_lov_user_md_v3(v3);
2444 if (size > sizeof(*v3))
2445 lustre_swab_lov_user_md_objects(v3->lmm_objects,
2448 CERROR("Invalid magic %#x\n", v1->lmm_magic);
2452 EXPORT_SYMBOL(lustre_swab_lov_comp_md_v1);
2454 void lustre_swab_lov_user_md_objects(struct lov_user_ost_data *lod,
2460 for (i = 0; i < stripe_count; i++) {
2461 lustre_swab_ost_id(&(lod[i].l_ost_oi));
2462 __swab32s(&(lod[i].l_ost_gen));
2463 __swab32s(&(lod[i].l_ost_idx));
2467 EXPORT_SYMBOL(lustre_swab_lov_user_md_objects);
2469 void lustre_swab_lov_user_md(struct lov_user_md *lum, size_t size)
2471 struct lov_user_md_v1 *v1;
2472 struct lov_user_md_v3 *v3;
2473 struct lov_foreign_md *lfm;
2477 CDEBUG(D_IOCTL, "swabbing lov_user_md\n");
2478 switch (lum->lmm_magic) {
2479 case __swab32(LOV_MAGIC_V1):
2480 case LOV_USER_MAGIC_V1:
2482 v1 = (struct lov_user_md_v1 *)lum;
2483 stripe_count = v1->lmm_stripe_count;
2485 if (lum->lmm_magic != LOV_USER_MAGIC_V1)
2486 __swab16s(&stripe_count);
2488 lustre_swab_lov_user_md_v1(v1);
2489 if (size > sizeof(*v1))
2490 lustre_swab_lov_user_md_objects(v1->lmm_objects,
2495 case __swab32(LOV_MAGIC_V3):
2496 case LOV_USER_MAGIC_V3:
2498 v3 = (struct lov_user_md_v3 *)lum;
2499 stripe_count = v3->lmm_stripe_count;
2501 if (lum->lmm_magic != LOV_USER_MAGIC_V3)
2502 __swab16s(&stripe_count);
2504 lustre_swab_lov_user_md_v3(v3);
2505 if (size > sizeof(*v3))
2506 lustre_swab_lov_user_md_objects(v3->lmm_objects,
2510 case __swab32(LOV_USER_MAGIC_SPECIFIC):
2511 case LOV_USER_MAGIC_SPECIFIC:
2513 v3 = (struct lov_user_md_v3 *)lum;
2514 stripe_count = v3->lmm_stripe_count;
2516 if (lum->lmm_magic != LOV_USER_MAGIC_SPECIFIC)
2517 __swab16s(&stripe_count);
2519 lustre_swab_lov_user_md_v3(v3);
2520 lustre_swab_lov_user_md_objects(v3->lmm_objects, stripe_count);
2523 case __swab32(LOV_MAGIC_COMP_V1):
2524 case LOV_USER_MAGIC_COMP_V1:
2525 lustre_swab_lov_comp_md_v1((struct lov_comp_md_v1 *)lum);
2527 case __swab32(LOV_MAGIC_FOREIGN):
2528 case LOV_USER_MAGIC_FOREIGN:
2530 lfm = (struct lov_foreign_md *)lum;
2531 __swab32s(&lfm->lfm_magic);
2532 __swab32s(&lfm->lfm_length);
2533 __swab32s(&lfm->lfm_type);
2534 __swab32s(&lfm->lfm_flags);
2538 CDEBUG(D_IOCTL, "Invalid LOV magic %08x\n", lum->lmm_magic);
2541 EXPORT_SYMBOL(lustre_swab_lov_user_md);
2543 void lustre_swab_lov_mds_md(struct lov_mds_md *lmm)
2546 CDEBUG(D_IOCTL, "swabbing lov_mds_md\n");
2547 __swab32s(&lmm->lmm_magic);
2548 __swab32s(&lmm->lmm_pattern);
2549 lustre_swab_lmm_oi(&lmm->lmm_oi);
2550 __swab32s(&lmm->lmm_stripe_size);
2551 __swab16s(&lmm->lmm_stripe_count);
2552 __swab16s(&lmm->lmm_layout_gen);
2555 EXPORT_SYMBOL(lustre_swab_lov_mds_md);
2557 void lustre_swab_ldlm_res_id(struct ldlm_res_id *id)
2561 for (i = 0; i < RES_NAME_SIZE; i++)
2562 __swab64s(&id->name[i]);
2565 void lustre_swab_ldlm_policy_data(union ldlm_wire_policy_data *d)
2567 /* the lock data is a union and the first two fields are always an
2568 * extent so it's ok to process an LDLM_EXTENT and LDLM_FLOCK lock
2569 * data the same way.
2571 __swab64s(&d->l_extent.start);
2572 __swab64s(&d->l_extent.end);
2573 __swab64s(&d->l_extent.gid);
2574 __swab64s(&d->l_flock.lfw_owner);
2575 __swab32s(&d->l_flock.lfw_pid);
2578 void lustre_swab_ldlm_intent(struct ldlm_intent *i)
2583 void lustre_swab_ldlm_resource_desc(struct ldlm_resource_desc *r)
2585 __swab32s(&r->lr_type);
2586 BUILD_BUG_ON(offsetof(typeof(*r), lr_pad) == 0);
2587 lustre_swab_ldlm_res_id(&r->lr_name);
2590 void lustre_swab_ldlm_lock_desc(struct ldlm_lock_desc *l)
2592 lustre_swab_ldlm_resource_desc(&l->l_resource);
2593 __swab32s(&l->l_req_mode);
2594 __swab32s(&l->l_granted_mode);
2595 lustre_swab_ldlm_policy_data(&l->l_policy_data);
2598 void lustre_swab_ldlm_request(struct ldlm_request *rq)
2600 __swab32s(&rq->lock_flags);
2601 lustre_swab_ldlm_lock_desc(&rq->lock_desc);
2602 __swab32s(&rq->lock_count);
2603 /* lock_handle[] opaque */
2606 void lustre_swab_ldlm_reply(struct ldlm_reply *r)
2608 __swab32s(&r->lock_flags);
2609 BUILD_BUG_ON(offsetof(typeof(*r), lock_padding) == 0);
2610 lustre_swab_ldlm_lock_desc(&r->lock_desc);
2611 /* lock_handle opaque */
2612 __swab64s(&r->lock_policy_res1);
2613 __swab64s(&r->lock_policy_res2);
2616 void lustre_swab_quota_body(struct quota_body *b)
2618 lustre_swab_lu_fid(&b->qb_fid);
2619 lustre_swab_lu_fid((struct lu_fid *)&b->qb_id);
2620 __swab32s(&b->qb_flags);
2621 __swab64s(&b->qb_count);
2622 __swab64s(&b->qb_usage);
2623 __swab64s(&b->qb_slv_ver);
2626 /* Dump functions */
2627 void dump_ioo(struct obd_ioobj *ioo)
2630 "obd_ioobj: ioo_oid="DOSTID", ioo_max_brw=%#x, "
2631 "ioo_bufct=%d\n", POSTID(&ioo->ioo_oid), ioo->ioo_max_brw,
2635 void dump_rniobuf(struct niobuf_remote *nb)
2637 CDEBUG(D_RPCTRACE, "niobuf_remote: offset=%llu, len=%d, flags=%x\n",
2638 nb->rnb_offset, nb->rnb_len, nb->rnb_flags);
2641 void dump_obdo(struct obdo *oa)
2643 u64 valid = oa->o_valid;
2645 CDEBUG(D_RPCTRACE, "obdo: o_valid = %#llx\n", valid);
2646 if (valid & OBD_MD_FLID)
2647 CDEBUG(D_RPCTRACE, "obdo: id = "DOSTID"\n", POSTID(&oa->o_oi));
2648 if (valid & OBD_MD_FLFID)
2649 CDEBUG(D_RPCTRACE, "obdo: o_parent_seq = %#llx\n",
2651 if (valid & OBD_MD_FLSIZE)
2652 CDEBUG(D_RPCTRACE, "obdo: o_size = %lld\n", oa->o_size);
2653 if (valid & OBD_MD_FLMTIME)
2654 CDEBUG(D_RPCTRACE, "obdo: o_mtime = %lld\n", oa->o_mtime);
2655 if (valid & OBD_MD_FLATIME)
2656 CDEBUG(D_RPCTRACE, "obdo: o_atime = %lld\n", oa->o_atime);
2657 if (valid & OBD_MD_FLCTIME)
2658 CDEBUG(D_RPCTRACE, "obdo: o_ctime = %lld\n", oa->o_ctime);
2659 if (valid & OBD_MD_FLBLOCKS) /* allocation of space */
2660 CDEBUG(D_RPCTRACE, "obdo: o_blocks = %lld\n", oa->o_blocks);
2661 if (valid & OBD_MD_FLGRANT)
2662 CDEBUG(D_RPCTRACE, "obdo: o_grant = %lld\n", oa->o_grant);
2663 if (valid & OBD_MD_FLBLKSZ)
2664 CDEBUG(D_RPCTRACE, "obdo: o_blksize = %d\n", oa->o_blksize);
2665 if (valid & (OBD_MD_FLTYPE | OBD_MD_FLMODE))
2666 CDEBUG(D_RPCTRACE, "obdo: o_mode = %o\n",
2667 oa->o_mode & ((valid & OBD_MD_FLTYPE ? S_IFMT : 0) |
2668 (valid & OBD_MD_FLMODE ? ~S_IFMT : 0)));
2669 if (valid & OBD_MD_FLUID)
2670 CDEBUG(D_RPCTRACE, "obdo: o_uid = %u\n", oa->o_uid);
2671 if (valid & OBD_MD_FLUID)
2672 CDEBUG(D_RPCTRACE, "obdo: o_uid_h = %u\n", oa->o_uid_h);
2673 if (valid & OBD_MD_FLGID)
2674 CDEBUG(D_RPCTRACE, "obdo: o_gid = %u\n", oa->o_gid);
2675 if (valid & OBD_MD_FLGID)
2676 CDEBUG(D_RPCTRACE, "obdo: o_gid_h = %u\n", oa->o_gid_h);
2677 if (valid & OBD_MD_FLFLAGS)
2678 CDEBUG(D_RPCTRACE, "obdo: o_flags = %x\n", oa->o_flags);
2679 if (valid & OBD_MD_FLNLINK)
2680 CDEBUG(D_RPCTRACE, "obdo: o_nlink = %u\n", oa->o_nlink);
2681 else if (valid & OBD_MD_FLCKSUM)
2682 CDEBUG(D_RPCTRACE, "obdo: o_checksum (o_nlink) = %u\n",
2684 if (valid & OBD_MD_FLPARENT)
2685 CDEBUG(D_RPCTRACE, "obdo: o_parent_oid = %x\n",
2687 if (valid & OBD_MD_FLFID) {
2688 CDEBUG(D_RPCTRACE, "obdo: o_stripe_idx = %u\n",
2690 CDEBUG(D_RPCTRACE, "obdo: o_parent_ver = %x\n",
2693 if (valid & OBD_MD_FLHANDLE)
2694 CDEBUG(D_RPCTRACE, "obdo: o_handle = %lld\n",
2695 oa->o_handle.cookie);
2698 void dump_ost_body(struct ost_body *ob)
2703 void dump_rcs(__u32 *rc)
2705 CDEBUG(D_RPCTRACE, "rmf_rcs: %d\n", *rc);
2708 static inline int req_ptlrpc_body_swabbed(struct ptlrpc_request *req)
2710 LASSERT(req->rq_reqmsg);
2712 switch (req->rq_reqmsg->lm_magic) {
2713 case LUSTRE_MSG_MAGIC_V2:
2714 return lustre_req_swabbed(req, MSG_PTLRPC_BODY_OFF);
2716 CERROR("bad lustre msg magic: %#08X\n",
2717 req->rq_reqmsg->lm_magic);
2722 static inline int rep_ptlrpc_body_swabbed(struct ptlrpc_request *req)
2724 if (unlikely(!req->rq_repmsg))
2727 switch (req->rq_repmsg->lm_magic) {
2728 case LUSTRE_MSG_MAGIC_V2:
2729 return lustre_rep_swabbed(req, MSG_PTLRPC_BODY_OFF);
2731 /* uninitialized yet */
2736 void _debug_req(struct ptlrpc_request *req,
2737 struct libcfs_debug_msg_data *msgdata, const char *fmt, ...)
2739 bool req_ok = req->rq_reqmsg != NULL;
2740 bool rep_ok = false;
2741 lnet_nid_t nid = LNET_NID_ANY;
2742 struct va_format vaf;
2745 int rep_status = -1;
2747 spin_lock(&req->rq_early_free_lock);
2751 if (ptlrpc_req_need_swab(req)) {
2752 req_ok = req_ok && req_ptlrpc_body_swabbed(req);
2753 rep_ok = rep_ok && rep_ptlrpc_body_swabbed(req);
2757 rep_flags = lustre_msg_get_flags(req->rq_repmsg);
2758 rep_status = lustre_msg_get_status(req->rq_repmsg);
2760 spin_unlock(&req->rq_early_free_lock);
2762 if (req->rq_import && req->rq_import->imp_connection)
2763 nid = req->rq_import->imp_connection->c_peer.nid;
2764 else if (req->rq_export && req->rq_export->exp_connection)
2765 nid = req->rq_export->exp_connection->c_peer.nid;
2767 va_start(args, fmt);
2770 libcfs_debug_msg(msgdata,
2771 "%pV req@%p x%llu/t%lld(%lld) o%d->%s@%s:%d/%d lens %d/%d e %d to %lld dl %lld ref %d fl " REQ_FLAGS_FMT "/%x/%x rc %d/%d job:'%s'\n",
2773 req, req->rq_xid, req->rq_transno,
2774 req_ok ? lustre_msg_get_transno(req->rq_reqmsg) : 0,
2775 req_ok ? lustre_msg_get_opc(req->rq_reqmsg) : -1,
2777 req->rq_import->imp_obd->obd_name :
2779 req->rq_export->exp_client_uuid.uuid :
2781 libcfs_nid2str(nid),
2782 req->rq_request_portal, req->rq_reply_portal,
2783 req->rq_reqlen, req->rq_replen,
2784 req->rq_early_count, (s64)req->rq_timedout,
2785 (s64)req->rq_deadline,
2786 atomic_read(&req->rq_refcount),
2787 DEBUG_REQ_FLAGS(req),
2788 req_ok ? lustre_msg_get_flags(req->rq_reqmsg) : -1,
2789 rep_flags, req->rq_status, rep_status,
2790 req_ok ? lustre_msg_get_jobid(req->rq_reqmsg) ?: ""
2794 EXPORT_SYMBOL(_debug_req);
2796 void lustre_swab_lustre_capa(struct lustre_capa *c)
2798 lustre_swab_lu_fid(&c->lc_fid);
2799 __swab64s(&c->lc_opc);
2800 __swab64s(&c->lc_uid);
2801 __swab64s(&c->lc_gid);
2802 __swab32s(&c->lc_flags);
2803 __swab32s(&c->lc_keyid);
2804 __swab32s(&c->lc_timeout);
2805 __swab32s(&c->lc_expiry);
2808 void lustre_swab_lustre_capa_key(struct lustre_capa_key *k)
2810 __swab64s(&k->lk_seq);
2811 __swab32s(&k->lk_keyid);
2812 BUILD_BUG_ON(offsetof(typeof(*k), lk_padding) == 0);
2815 void lustre_swab_hsm_user_state(struct hsm_user_state *state)
2817 __swab32s(&state->hus_states);
2818 __swab32s(&state->hus_archive_id);
2821 void lustre_swab_hsm_state_set(struct hsm_state_set *hss)
2823 __swab32s(&hss->hss_valid);
2824 __swab64s(&hss->hss_setmask);
2825 __swab64s(&hss->hss_clearmask);
2826 __swab32s(&hss->hss_archive_id);
2829 static void lustre_swab_hsm_extent(struct hsm_extent *extent)
2831 __swab64s(&extent->offset);
2832 __swab64s(&extent->length);
2835 void lustre_swab_hsm_current_action(struct hsm_current_action *action)
2837 __swab32s(&action->hca_state);
2838 __swab32s(&action->hca_action);
2839 lustre_swab_hsm_extent(&action->hca_location);
2842 void lustre_swab_hsm_user_item(struct hsm_user_item *hui)
2844 lustre_swab_lu_fid(&hui->hui_fid);
2845 lustre_swab_hsm_extent(&hui->hui_extent);
2848 void lustre_swab_lu_extent(struct lu_extent *le)
2850 __swab64s(&le->e_start);
2851 __swab64s(&le->e_end);
2854 void lustre_swab_layout_intent(struct layout_intent *li)
2856 __swab32s(&li->li_opc);
2857 __swab32s(&li->li_flags);
2858 lustre_swab_lu_extent(&li->li_extent);
2861 void lustre_swab_hsm_progress_kernel(struct hsm_progress_kernel *hpk)
2863 lustre_swab_lu_fid(&hpk->hpk_fid);
2864 __swab64s(&hpk->hpk_cookie);
2865 __swab64s(&hpk->hpk_extent.offset);
2866 __swab64s(&hpk->hpk_extent.length);
2867 __swab16s(&hpk->hpk_flags);
2868 __swab16s(&hpk->hpk_errval);
2871 void lustre_swab_hsm_request(struct hsm_request *hr)
2873 __swab32s(&hr->hr_action);
2874 __swab32s(&hr->hr_archive_id);
2875 __swab64s(&hr->hr_flags);
2876 __swab32s(&hr->hr_itemcount);
2877 __swab32s(&hr->hr_data_len);
2880 void lustre_swab_object_update(struct object_update *ou)
2882 struct object_update_param *param;
2885 __swab16s(&ou->ou_type);
2886 __swab16s(&ou->ou_params_count);
2887 __swab32s(&ou->ou_result_size);
2888 __swab32s(&ou->ou_flags);
2889 __swab32s(&ou->ou_padding1);
2890 __swab64s(&ou->ou_batchid);
2891 lustre_swab_lu_fid(&ou->ou_fid);
2892 param = &ou->ou_params[0];
2893 for (i = 0; i < ou->ou_params_count; i++) {
2894 __swab16s(¶m->oup_len);
2895 __swab16s(¶m->oup_padding);
2896 __swab32s(¶m->oup_padding2);
2897 param = (struct object_update_param *)((char *)param +
2898 object_update_param_size(param));
2902 void lustre_swab_object_update_request(struct object_update_request *our)
2905 __swab32s(&our->ourq_magic);
2906 __swab16s(&our->ourq_count);
2907 __swab16s(&our->ourq_padding);
2908 for (i = 0; i < our->ourq_count; i++) {
2909 struct object_update *ou;
2911 ou = object_update_request_get(our, i, NULL);
2914 lustre_swab_object_update(ou);
2918 void lustre_swab_object_update_result(struct object_update_result *our)
2920 __swab32s(&our->our_rc);
2921 __swab16s(&our->our_datalen);
2922 __swab16s(&our->our_padding);
2925 void lustre_swab_object_update_reply(struct object_update_reply *our)
2929 __swab32s(&our->ourp_magic);
2930 __swab16s(&our->ourp_count);
2931 __swab16s(&our->ourp_padding);
2932 for (i = 0; i < our->ourp_count; i++) {
2933 struct object_update_result *ourp;
2935 __swab16s(&our->ourp_lens[i]);
2936 ourp = object_update_result_get(our, i, NULL);
2939 lustre_swab_object_update_result(ourp);
2943 void lustre_swab_out_update_header(struct out_update_header *ouh)
2945 __swab32s(&ouh->ouh_magic);
2946 __swab32s(&ouh->ouh_count);
2947 __swab32s(&ouh->ouh_inline_length);
2948 __swab32s(&ouh->ouh_reply_size);
2950 EXPORT_SYMBOL(lustre_swab_out_update_header);
2952 void lustre_swab_out_update_buffer(struct out_update_buffer *oub)
2954 __swab32s(&oub->oub_size);
2955 __swab32s(&oub->oub_padding);
2957 EXPORT_SYMBOL(lustre_swab_out_update_buffer);
2959 void lustre_swab_swap_layouts(struct mdc_swap_layouts *msl)
2961 __swab64s(&msl->msl_flags);
2964 void lustre_swab_close_data(struct close_data *cd)
2966 lustre_swab_lu_fid(&cd->cd_fid);
2967 __swab64s(&cd->cd_data_version);
2970 void lustre_swab_close_data_resync_done(struct close_data_resync_done *resync)
2974 __swab32s(&resync->resync_count);
2975 /* after swab, resync_count must in CPU endian */
2976 if (resync->resync_count <= INLINE_RESYNC_ARRAY_SIZE) {
2977 for (i = 0; i < resync->resync_count; i++)
2978 __swab32s(&resync->resync_ids_inline[i]);
2981 EXPORT_SYMBOL(lustre_swab_close_data_resync_done);
2983 void lustre_swab_lfsck_request(struct lfsck_request *lr)
2985 __swab32s(&lr->lr_event);
2986 __swab32s(&lr->lr_index);
2987 __swab32s(&lr->lr_flags);
2988 __swab32s(&lr->lr_valid);
2989 __swab32s(&lr->lr_speed);
2990 __swab16s(&lr->lr_version);
2991 __swab16s(&lr->lr_active);
2992 __swab16s(&lr->lr_param);
2993 __swab16s(&lr->lr_async_windows);
2994 __swab32s(&lr->lr_flags);
2995 lustre_swab_lu_fid(&lr->lr_fid);
2996 lustre_swab_lu_fid(&lr->lr_fid2);
2997 __swab32s(&lr->lr_comp_id);
2998 BUILD_BUG_ON(offsetof(typeof(*lr), lr_padding_0) == 0);
2999 BUILD_BUG_ON(offsetof(typeof(*lr), lr_padding_1) == 0);
3000 BUILD_BUG_ON(offsetof(typeof(*lr), lr_padding_2) == 0);
3001 BUILD_BUG_ON(offsetof(typeof(*lr), lr_padding_3) == 0);
3004 void lustre_swab_lfsck_reply(struct lfsck_reply *lr)
3006 __swab32s(&lr->lr_status);
3007 BUILD_BUG_ON(offsetof(typeof(*lr), lr_padding_1) == 0);
3008 __swab64s(&lr->lr_repaired);
3011 static void lustre_swab_orphan_rec(struct lu_orphan_rec *rec)
3013 lustre_swab_lu_fid(&rec->lor_fid);
3014 __swab32s(&rec->lor_uid);
3015 __swab32s(&rec->lor_gid);
3018 void lustre_swab_orphan_ent(struct lu_orphan_ent *ent)
3020 lustre_swab_lu_fid(&ent->loe_key);
3021 lustre_swab_orphan_rec(&ent->loe_rec);
3023 EXPORT_SYMBOL(lustre_swab_orphan_ent);
3025 void lustre_swab_orphan_ent_v2(struct lu_orphan_ent_v2 *ent)
3027 lustre_swab_lu_fid(&ent->loe_key);
3028 lustre_swab_orphan_rec(&ent->loe_rec.lor_rec);
3029 lustre_swab_ost_layout(&ent->loe_rec.lor_layout);
3030 BUILD_BUG_ON(offsetof(typeof(ent->loe_rec), lor_padding) == 0);
3032 EXPORT_SYMBOL(lustre_swab_orphan_ent_v2);
3034 void lustre_swab_orphan_ent_v3(struct lu_orphan_ent_v3 *ent)
3036 lustre_swab_lu_fid(&ent->loe_key);
3037 lustre_swab_orphan_rec(&ent->loe_rec.lor_rec);
3038 lustre_swab_ost_layout(&ent->loe_rec.lor_layout);
3039 __swab32s(&ent->loe_rec.lor_layout_version);
3040 __swab32s(&ent->loe_rec.lor_range);
3041 BUILD_BUG_ON(offsetof(typeof(ent->loe_rec), lor_padding_1) == 0);
3042 BUILD_BUG_ON(offsetof(typeof(ent->loe_rec), lor_padding_2) == 0);
3044 EXPORT_SYMBOL(lustre_swab_orphan_ent_v3);
3046 void lustre_swab_ladvise(struct lu_ladvise *ladvise)
3048 __swab16s(&ladvise->lla_advice);
3049 __swab16s(&ladvise->lla_value1);
3050 __swab32s(&ladvise->lla_value2);
3051 __swab64s(&ladvise->lla_start);
3052 __swab64s(&ladvise->lla_end);
3053 __swab32s(&ladvise->lla_value3);
3054 __swab32s(&ladvise->lla_value4);
3056 EXPORT_SYMBOL(lustre_swab_ladvise);
3058 void lustre_swab_ladvise_hdr(struct ladvise_hdr *ladvise_hdr)
3060 __swab32s(&ladvise_hdr->lah_magic);
3061 __swab32s(&ladvise_hdr->lah_count);
3062 __swab64s(&ladvise_hdr->lah_flags);
3063 __swab32s(&ladvise_hdr->lah_value1);
3064 __swab32s(&ladvise_hdr->lah_value2);
3065 __swab64s(&ladvise_hdr->lah_value3);
3067 EXPORT_SYMBOL(lustre_swab_ladvise_hdr);