Whamcloud - gitweb
LU-3963 obdclass: convert to linux list api
[fs/lustre-release.git] / lustre / include / lustre_net.h
index c5f5d86..70600b9 100644 (file)
@@ -27,7 +27,7 @@
  * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
  * Use is subject to license terms.
  *
- * Copyright (c) 2010, 2012, Intel Corporation.
+ * Copyright (c) 2010, 2013, Intel Corporation.
  */
 /*
  * This file is part of Lustre, http://www.lustre.org/
  */
 #define PTLRPC_MAX_BRW_BITS    (LNET_MTU_BITS + PTLRPC_BULK_OPS_BITS)
 #define PTLRPC_MAX_BRW_SIZE    (1 << PTLRPC_MAX_BRW_BITS)
-#define PTLRPC_MAX_BRW_PAGES   (PTLRPC_MAX_BRW_SIZE >> CFS_PAGE_SHIFT)
+#define PTLRPC_MAX_BRW_PAGES   (PTLRPC_MAX_BRW_SIZE >> PAGE_CACHE_SHIFT)
 
 #define ONE_MB_BRW_SIZE                (1 << LNET_MTU_BITS)
 #define MD_MAX_BRW_SIZE                (1 << LNET_MTU_BITS)
-#define MD_MAX_BRW_PAGES       (MD_MAX_BRW_SIZE >> CFS_PAGE_SHIFT)
+#define MD_MAX_BRW_PAGES       (MD_MAX_BRW_SIZE >> PAGE_CACHE_SHIFT)
 #define DT_MAX_BRW_SIZE                PTLRPC_MAX_BRW_SIZE
-#define DT_MAX_BRW_PAGES       (DT_MAX_BRW_SIZE >> CFS_PAGE_SHIFT)
+#define DT_MAX_BRW_PAGES       (DT_MAX_BRW_SIZE >> PAGE_CACHE_SHIFT)
 #define OFD_MAX_BRW_SIZE       (1 << LNET_MTU_BITS)
 
 /* When PAGE_SIZE is a constant, we can check our arithmetic here with cpp! */
 # if ((PTLRPC_MAX_BRW_PAGES & (PTLRPC_MAX_BRW_PAGES - 1)) != 0)
 #  error "PTLRPC_MAX_BRW_PAGES isn't a power of two"
 # endif
-# if (PTLRPC_MAX_BRW_SIZE != (PTLRPC_MAX_BRW_PAGES * CFS_PAGE_SIZE))
-#  error "PTLRPC_MAX_BRW_SIZE isn't PTLRPC_MAX_BRW_PAGES * CFS_PAGE_SIZE"
+# if (PTLRPC_MAX_BRW_SIZE != (PTLRPC_MAX_BRW_PAGES * PAGE_CACHE_SIZE))
+#  error "PTLRPC_MAX_BRW_SIZE isn't PTLRPC_MAX_BRW_PAGES * PAGE_CACHE_SIZE"
 # endif
 # if (PTLRPC_MAX_BRW_SIZE > LNET_MTU * PTLRPC_BULK_OPS_COUNT)
 #  error "PTLRPC_MAX_BRW_SIZE too big"
 #define LDLM_THR_FACTOR                8
 #define LDLM_NTHRS_INIT                PTLRPC_NTHRS_INIT
 #define LDLM_NTHRS_BASE                24
-#define LDLM_NTHRS_MAX         (cfs_num_online_cpus() == 1 ? 64 : 128)
+#define LDLM_NTHRS_MAX         (num_online_cpus() == 1 ? 64 : 128)
 
 #define LDLM_BL_THREADS   LDLM_NTHRS_AUTO_INIT
 #define LDLM_CLIENT_NBUFS 1
  * include linkea (4K maxim), together with other updates, we set it to 9K:
  * lustre_msg + ptlrpc_body + UPDATE_BUF_SIZE (8K)
  */
-#define MDS_OUT_MAXREQSIZE     (9 * 1024)
-#define MDS_OUT_MAXREPSIZE     MDS_MAXREPSIZE
+#define OUT_MAXREQSIZE (9 * 1024)
+#define OUT_MAXREPSIZE MDS_MAXREPSIZE
 
 /** MDS_BUFSIZE = max_reqsize (w/o LOV EA) + max sptlrpc payload size */
 #define MDS_BUFSIZE            max(MDS_MAXREQSIZE + SPTLRPC_MAX_PAYLOAD, \
                                    160 * 1024)
 
 /**
- * MDS_OUT_BUFSIZE = max_out_reqsize + max sptlrpc payload (~1K) which is
+ * OUT_BUFSIZE = max_out_reqsize + max sptlrpc payload (~1K) which is
  * about 10K, for the same reason as MDS_REG_BUFSIZE, we also give some
  * extra bytes to each request buffer to improve buffer utilization rate.
   */
-#define MDS_OUT_BUFSIZE                max(MDS_OUT_MAXREQSIZE + SPTLRPC_MAX_PAYLOAD, \
+#define OUT_BUFSIZE            max(OUT_MAXREQSIZE + SPTLRPC_MAX_PAYLOAD, \
                                    24 * 1024)
 
 /** FLD_MAXREQSIZE == lustre_msg + __u32 padding + ptlrpc_body + opc */
   */
  /* depress threads factor for VM with small memory size */
 #define OSS_THR_FACTOR         min_t(int, 8, \
-                               CFS_NUM_CACHEPAGES >> (28 - CFS_PAGE_SHIFT))
+                               NUM_CACHEPAGES >> (28 - PAGE_CACHE_SHIFT))
 #define OSS_NTHRS_INIT         (PTLRPC_NTHRS_INIT + 1)
 #define OSS_NTHRS_BASE         64
 #define OSS_NTHRS_MAX          512
 /**
  * FIEMAP request can be 4K+ for now
  */
-#define OST_MAXREQSIZE         (5 * 1024)
+#define OST_MAXREQSIZE         (16 * 1024)
 #define OST_IO_MAXREQSIZE      max_t(int, OST_MAXREQSIZE, \
                                (((_OST_MAXREQSIZE_SUM - 1) | (1024 - 1)) + 1))
 
  * Structure to single define portal connection.
  */
 struct ptlrpc_connection {
-        /** linkage for connections hash table */
-        cfs_hlist_node_t        c_hash;
-        /** Our own lnet nid for this connection */
-        lnet_nid_t              c_self;
-        /** Remote side nid for this connection */
-        lnet_process_id_t       c_peer;
-        /** UUID of the other side */
-        struct obd_uuid         c_remote_uuid;
-        /** reference counter for this connection */
-        cfs_atomic_t            c_refcount;
+       /** linkage for connections hash table */
+       struct hlist_node        c_hash;
+       /** Our own lnet nid for this connection */
+       lnet_nid_t              c_self;
+       /** Remote side nid for this connection */
+       lnet_process_id_t       c_peer;
+       /** UUID of the other side */
+       struct obd_uuid         c_remote_uuid;
+       /** reference counter for this connection */
+       atomic_t            c_refcount;
 };
 
 /** Client definition for PortalRPC */
@@ -574,26 +574,26 @@ typedef int (*set_producer_func)(struct ptlrpc_request_set *, void *);
  * returned.
  */
 struct ptlrpc_request_set {
-       cfs_atomic_t          set_refcount;
+       atomic_t                set_refcount;
        /** number of in queue requests */
-       cfs_atomic_t          set_new_count;
+       atomic_t                set_new_count;
        /** number of uncompleted requests */
-       cfs_atomic_t          set_remaining;
+       atomic_t                set_remaining;
        /** wait queue to wait on for request events */
-       cfs_waitq_t           set_waitq;
-       cfs_waitq_t          *set_wakeup_ptr;
+       wait_queue_head_t       set_waitq;
+       wait_queue_head_t      *set_wakeup_ptr;
        /** List of requests in the set */
-       cfs_list_t            set_requests;
+       struct list_head        set_requests;
        /**
         * List of completion callbacks to be called when the set is completed
         * This is only used if \a set_interpret is NULL.
         * Links struct ptlrpc_set_cbdata.
         */
-       cfs_list_t            set_cblist;
+       struct list_head        set_cblist;
        /** Completion callback, if only one. */
-       set_interpreter_func  set_interpret;
+       set_interpreter_func    set_interpret;
        /** opaq argument passed to completion \a set_interpret callback. */
-       void                 *set_arg;
+       void                    *set_arg;
        /**
         * Lock for \a set_new_requests manipulations
         * locked so that any old caller can communicate requests to
@@ -601,29 +601,29 @@ struct ptlrpc_request_set {
         */
        spinlock_t              set_new_req_lock;
        /** List of new yet unsent requests. Only used with ptlrpcd now. */
-       cfs_list_t            set_new_requests;
+       struct list_head        set_new_requests;
 
        /** rq_status of requests that have been freed already */
-       int                   set_rc;
+       int                     set_rc;
        /** Additional fields used by the flow control extension */
        /** Maximum number of RPCs in flight */
-       int                   set_max_inflight;
+       int                     set_max_inflight;
        /** Callback function used to generate RPCs */
-       set_producer_func     set_producer;
+       set_producer_func       set_producer;
        /** opaq argument passed to the producer callback */
-       void                 *set_producer_arg;
+       void                    *set_producer_arg;
 };
 
 /**
  * Description of a single ptrlrpc_set callback
  */
 struct ptlrpc_set_cbdata {
-        /** List linkage item */
-        cfs_list_t              psc_item;
-        /** Pointer to interpreting function */
-        set_interpreter_func    psc_interpret;
-        /** Opaq argument to pass to the callback */
-        void                   *psc_data;
+       /** List linkage item */
+       struct list_head        psc_item;
+       /** Pointer to interpreting function */
+       set_interpreter_func    psc_interpret;
+       /** Opaq argument to pass to the callback */
+       void                    *psc_data;
 };
 
 struct ptlrpc_bulk_desc;
@@ -650,20 +650,20 @@ struct ptlrpc_cb_id {
  * added to the state for replay/failover consistency guarantees.
  */
 struct ptlrpc_reply_state {
-        /** Callback description */
-        struct ptlrpc_cb_id    rs_cb_id;
-        /** Linkage for list of all reply states in a system */
-        cfs_list_t             rs_list;
-        /** Linkage for list of all reply states on same export */
-        cfs_list_t             rs_exp_list;
-        /** Linkage for list of all reply states for same obd */
-        cfs_list_t             rs_obd_list;
+       /** Callback description */
+       struct ptlrpc_cb_id     rs_cb_id;
+       /** Linkage for list of all reply states in a system */
+       struct list_head        rs_list;
+       /** Linkage for list of all reply states on same export */
+       struct list_head        rs_exp_list;
+       /** Linkage for list of all reply states for same obd */
+       struct list_head        rs_obd_list;
 #if RS_DEBUG
-        cfs_list_t             rs_debug_list;
+       struct list_head        rs_debug_list;
 #endif
-        /** A spinlock to protect the reply state flags */
+       /** A spinlock to protect the reply state flags */
        spinlock_t              rs_lock;
-        /** Reply state flags */
+       /** Reply state flags */
         unsigned long          rs_difficult:1;     /* ACK/commit stuff */
         unsigned long          rs_no_ack:1;    /* no ACK, even for
                                                   difficult requests */
@@ -683,16 +683,16 @@ struct ptlrpc_reply_state {
         __u64                  rs_transno;
         /** xid */
         __u64                  rs_xid;
-        struct obd_export     *rs_export;
+       struct obd_export     *rs_export;
        struct ptlrpc_service_part *rs_svcpt;
-        /** Lnet metadata handle for the reply */
-        lnet_handle_md_t       rs_md_h;
-        cfs_atomic_t           rs_refcount;
-
-        /** Context for the sevice thread */
-        struct ptlrpc_svc_ctx *rs_svc_ctx;
-        /** Reply buffer (actually sent to the client), encoded if needed */
-        struct lustre_msg     *rs_repbuf;       /* wrapper */
+       /** Lnet metadata handle for the reply */
+       lnet_handle_md_t       rs_md_h;
+       atomic_t               rs_refcount;
+
+       /** Context for the sevice thread */
+       struct ptlrpc_svc_ctx *rs_svc_ctx;
+       /** Reply buffer (actually sent to the client), encoded if needed */
+       struct lustre_msg     *rs_repbuf;       /* wrapper */
         /** Size of the reply buffer */
         int                    rs_repbuf_len;   /* wrapper buf length */
         /** Size of the reply message */
@@ -738,13 +738,13 @@ typedef int (*ptlrpc_interpterer_t)(const struct lu_env *env,
  */
 struct ptlrpc_request_pool {
        /** Locks the list */
-       spinlock_t prp_lock;
-        /** list of ptlrpc_request structs */
-        cfs_list_t prp_req_list;
-        /** Maximum message size that would fit into a rquest from this pool */
-        int prp_rq_size;
-        /** Function to allocate more requests for this pool */
-        void (*prp_populate)(struct ptlrpc_request_pool *, int);
+       spinlock_t              prp_lock;
+       /** list of ptlrpc_request structs */
+       struct list_head        prp_req_list;
+       /** Maximum message size that would fit into a rquest from this pool */
+       int                     prp_rq_size;
+       /** Function to allocate more requests for this pool */
+       void (*prp_populate)(struct ptlrpc_request_pool *, int);
 };
 
 struct lu_context;
@@ -787,6 +787,18 @@ enum ptlrpc_nrs_ctl {
 };
 
 /**
+ * ORR policy operations
+ */
+enum nrs_ctl_orr {
+       NRS_CTL_ORR_RD_QUANTUM = PTLRPC_NRS_CTL_1ST_POL_SPEC,
+       NRS_CTL_ORR_WR_QUANTUM,
+       NRS_CTL_ORR_RD_OFF_TYPE,
+       NRS_CTL_ORR_WR_OFF_TYPE,
+       NRS_CTL_ORR_RD_SUPP_REQ,
+       NRS_CTL_ORR_WR_SUPP_REQ,
+};
+
+/**
  * NRS policy operations.
  *
  * These determine the behaviour of a policy, and are called in response to
@@ -810,10 +822,12 @@ struct ptlrpc_nrs_pol_ops {
         * initialize their resources here; this operation is optional.
         *
         * \param[in,out] policy The policy being started
+        * \param[in,out] arg A generic char buffer
         *
         * \see nrs_policy_start_locked()
         */
-       int     (*op_policy_start) (struct ptlrpc_nrs_policy *policy);
+       int     (*op_policy_start) (struct ptlrpc_nrs_policy *policy,
+                                   char *arg);
        /**
         * Called when deactivating a policy via lprocfs; policies deallocate
         * their resources here; this operation is optional
@@ -949,7 +963,7 @@ struct ptlrpc_nrs_pol_ops {
         *                       \a nrq
         * \param[in,out] nrq    The request
         *
-        * \pre spin_is_locked(&svcpt->scp_req_lock)
+        * \pre assert_spin_locked(&svcpt->scp_req_lock)
         *
         * \see ptlrpc_nrs_req_stop_nolock()
         */
@@ -1044,7 +1058,7 @@ struct ptlrpc_nrs {
        /**
         * List of registered policies
         */
-       cfs_list_t                      nrs_policy_list;
+       struct list_head                nrs_policy_list;
        /**
         * List of policies with queued requests. Policies that have any
         * outstanding requests are queued here, and this list is queried
@@ -1053,7 +1067,7 @@ struct ptlrpc_nrs {
         * point transition away from the
         * ptlrpc_nrs_pol_state::NRS_POL_STATE_STARTED state are drained.
         */
-       cfs_list_t                      nrs_policy_queued;
+       struct list_head                nrs_policy_queued;
        /**
         * Service partition for this NRS head
         */
@@ -1091,6 +1105,10 @@ struct ptlrpc_nrs {
         * unregistration
         */
        unsigned                        nrs_stopping:1;
+       /**
+        * NRS policy is throttling reqeust
+        */
+       unsigned                        nrs_throttling:1;
 };
 
 #define NRS_POL_NAME_MAX               16
@@ -1134,7 +1152,7 @@ struct ptlrpc_nrs_pol_conf {
         * different module to the one the NRS framework is held within
         * (currently ptlrpc), should set this field to THIS_MODULE.
         */
-       cfs_module_t                      *nc_owner;
+       struct module                     *nc_owner;
        /**
         * Policy registration flags; a bitmast of \e nrs_policy_flags
         */
@@ -1155,7 +1173,7 @@ struct ptlrpc_nrs_pol_desc {
        /**
         * Link into nrs_core::nrs_policies
         */
-       cfs_list_t                              pd_list;
+       struct list_head                        pd_list;
        /**
         * NRS operations for this policy
         */
@@ -1209,7 +1227,7 @@ struct ptlrpc_nrs_pol_desc {
         *   then unregistration and lprocfs operations will be properly
         *   serialized.
         */
-       cfs_module_t                           *pd_owner;
+       struct module                          *pd_owner;
        /**
         * Bitmask of \e nrs_policy_flags
         */
@@ -1217,7 +1235,7 @@ struct ptlrpc_nrs_pol_desc {
        /**
         * # of references on this descriptor
         */
-       cfs_atomic_t                            pd_refs;
+       atomic_t                                pd_refs;
 };
 
 /**
@@ -1291,12 +1309,12 @@ struct ptlrpc_nrs_policy {
         * Linkage into the NRS head's list of policies,
         * ptlrpc_nrs:nrs_policy_list
         */
-       cfs_list_t                      pol_list;
+       struct list_head                pol_list;
        /**
         * Linkage into the NRS head's list of policies with enqueued
         * requests ptlrpc_nrs:nrs_policy_queued
         */
-       cfs_list_t                      pol_list_queued;
+       struct list_head                pol_list_queued;
        /**
         * Current state of this policy
         */
@@ -1398,7 +1416,7 @@ struct nrs_fifo_head {
        /**
         * List of queued requests.
         */
-       cfs_list_t                      fh_list;
+       struct list_head                fh_list;
        /**
         * For debugging purposes.
         */
@@ -1406,13 +1424,285 @@ struct nrs_fifo_head {
 };
 
 struct nrs_fifo_req {
-       cfs_list_t              fr_list;
+       struct list_head        fr_list;
        __u64                   fr_sequence;
 };
 
 /** @} fifo */
 
 /**
+ * \name CRR-N
+ *
+ * CRR-N, Client Round Robin over NIDs
+ * @{
+ */
+
+/**
+ * private data structure for CRR-N NRS
+ */
+struct nrs_crrn_net {
+       struct ptlrpc_nrs_resource      cn_res;
+       cfs_binheap_t                  *cn_binheap;
+       cfs_hash_t                     *cn_cli_hash;
+       /**
+        * Used when a new scheduling round commences, in order to synchronize
+        * all clients with the new round number.
+        */
+       __u64                           cn_round;
+       /**
+        * Determines the relevant ordering amongst request batches within a
+        * scheduling round.
+        */
+       __u64                           cn_sequence;
+       /**
+        * Round Robin quantum; the maximum number of RPCs that each request
+        * batch for each client can have in a scheduling round.
+        */
+       __u16                           cn_quantum;
+};
+
+/**
+ * Object representing a client in CRR-N, as identified by its NID
+ */
+struct nrs_crrn_client {
+       struct ptlrpc_nrs_resource      cc_res;
+       struct hlist_node               cc_hnode;
+       lnet_nid_t                      cc_nid;
+       /**
+        * The round number against which this client is currently scheduling
+        * requests.
+        */
+       __u64                           cc_round;
+       /**
+        * The sequence number used for requests scheduled by this client during
+        * the current round number.
+        */
+       __u64                           cc_sequence;
+       atomic_t                        cc_ref;
+       /**
+        * Round Robin quantum; the maximum number of RPCs the client is allowed
+        * to schedule in a single batch of each round.
+        */
+       __u16                           cc_quantum;
+       /**
+        * # of pending requests for this client, on all existing rounds
+        */
+       __u16                           cc_active;
+};
+
+/**
+ * CRR-N NRS request definition
+ */
+struct nrs_crrn_req {
+       /**
+        * Round number for this request; shared with all other requests in the
+        * same batch.
+        */
+       __u64                   cr_round;
+       /**
+        * Sequence number for this request; shared with all other requests in
+        * the same batch.
+        */
+       __u64                   cr_sequence;
+};
+
+/**
+ * CRR-N policy operations.
+ */
+enum nrs_ctl_crr {
+       /**
+        * Read the RR quantum size of a CRR-N policy.
+        */
+       NRS_CTL_CRRN_RD_QUANTUM = PTLRPC_NRS_CTL_1ST_POL_SPEC,
+       /**
+        * Write the RR quantum size of a CRR-N policy.
+        */
+       NRS_CTL_CRRN_WR_QUANTUM,
+};
+
+/** @} CRR-N */
+
+/**
+ * \name ORR/TRR
+ *
+ * ORR/TRR (Object-based Round Robin/Target-based Round Robin) NRS policies
+ * @{
+ */
+
+/**
+ * Lower and upper byte offsets of a brw RPC
+ */
+struct nrs_orr_req_range {
+       __u64           or_start;
+       __u64           or_end;
+};
+
+/**
+ * RPC types supported by the ORR/TRR policies
+ */
+enum nrs_orr_supp {
+       NOS_OST_READ  = (1 << 0),
+       NOS_OST_WRITE = (1 << 1),
+       NOS_OST_RW    = (NOS_OST_READ | NOS_OST_WRITE),
+       /**
+        * Default value for policies.
+        */
+       NOS_DFLT      = NOS_OST_READ
+};
+
+/**
+ * As unique keys for grouping RPCs together, we use the object's OST FID for
+ * the ORR policy, and the OST index for the TRR policy.
+ *
+ * XXX: We waste some space for TRR policy instances by using a union, but it
+ *     allows to consolidate some of the code between ORR and TRR, and these
+ *     policies will probably eventually merge into one anyway.
+ */
+struct nrs_orr_key {
+       union {
+               /** object FID for ORR */
+               struct lu_fid   ok_fid;
+               /** OST index for TRR */
+               __u32           ok_idx;
+       };
+};
+
+/**
+ * The largest base string for unique hash/slab object names is
+ * "nrs_orr_reg_", so 13 characters. We add 3 to this to be used for the CPT
+ * id number, so this _should_ be more than enough for the maximum number of
+ * CPTs on any system. If it does happen that this statement is incorrect,
+ * nrs_orr_genobjname() will inevitably yield a non-unique name and cause
+ * kmem_cache_create() to complain (on Linux), so the erroneous situation
+ * will hopefully not go unnoticed.
+ */
+#define NRS_ORR_OBJ_NAME_MAX   (sizeof("nrs_orr_reg_") + 3)
+
+/**
+ * private data structure for ORR and TRR NRS
+ */
+struct nrs_orr_data {
+       struct ptlrpc_nrs_resource      od_res;
+       cfs_binheap_t                  *od_binheap;
+       cfs_hash_t                     *od_obj_hash;
+       struct kmem_cache              *od_cache;
+       /**
+        * Used when a new scheduling round commences, in order to synchronize
+        * all object or OST batches with the new round number.
+        */
+       __u64                           od_round;
+       /**
+        * Determines the relevant ordering amongst request batches within a
+        * scheduling round.
+        */
+       __u64                           od_sequence;
+       /**
+        * RPC types that are currently supported.
+        */
+       enum nrs_orr_supp               od_supp;
+       /**
+        * Round Robin quantum; the maxium number of RPCs that each request
+        * batch for each object or OST can have in a scheduling round.
+        */
+       __u16                           od_quantum;
+       /**
+        * Whether to use physical disk offsets or logical file offsets.
+        */
+       bool                            od_physical;
+       /**
+        * XXX: We need to provide a persistently allocated string to hold
+        * unique object names for this policy, since in currently supported
+        * versions of Linux by Lustre, kmem_cache_create() just sets a pointer
+        * to the name string provided. kstrdup() is used in the version of
+        * kmeme_cache_create() in current Linux mainline, so we may be able to
+        * remove this in the future.
+        */
+       char                            od_objname[NRS_ORR_OBJ_NAME_MAX];
+};
+
+/**
+ * Represents a backend-fs object or OST in the ORR and TRR policies
+ * respectively
+ */
+struct nrs_orr_object {
+       struct ptlrpc_nrs_resource      oo_res;
+       struct hlist_node               oo_hnode;
+       /**
+        * The round number against which requests are being scheduled for this
+        * object or OST
+        */
+       __u64                           oo_round;
+       /**
+        * The sequence number used for requests scheduled for this object or
+        * OST during the current round number.
+        */
+       __u64                           oo_sequence;
+       /**
+        * The key of the object or OST for which this structure instance is
+        * scheduling RPCs
+        */
+       struct nrs_orr_key              oo_key;
+       long                            oo_ref;
+       /**
+        * Round Robin quantum; the maximum number of RPCs that are allowed to
+        * be scheduled for the object or OST in a single batch of each round.
+        */
+       __u16                           oo_quantum;
+       /**
+        * # of pending requests for this object or OST, on all existing rounds
+        */
+       __u16                           oo_active;
+};
+
+/**
+ * ORR/TRR NRS request definition
+ */
+struct nrs_orr_req {
+       /**
+        * The offset range this request covers
+        */
+       struct nrs_orr_req_range        or_range;
+       /**
+        * Round number for this request; shared with all other requests in the
+        * same batch.
+        */
+       __u64                           or_round;
+       /**
+        * Sequence number for this request; shared with all other requests in
+        * the same batch.
+        */
+       __u64                           or_sequence;
+       /**
+        * For debugging purposes.
+        */
+       struct nrs_orr_key              or_key;
+       /**
+        * An ORR policy instance has filled in request information while
+        * enqueueing the request on the service partition's regular NRS head.
+        */
+       unsigned int                    or_orr_set:1;
+       /**
+        * A TRR policy instance has filled in request information while
+        * enqueueing the request on the service partition's regular NRS head.
+        */
+       unsigned int                    or_trr_set:1;
+       /**
+        * Request offset ranges have been filled in with logical offset
+        * values.
+        */
+       unsigned int                    or_logical_set:1;
+       /**
+        * Request offset ranges have been filled in with physical offset
+        * values.
+        */
+       unsigned int                    or_physical_set:1;
+};
+
+/** @} ORR/TRR */
+
+#include <lustre_nrs_tbf.h>
+
+/**
  * NRS request
  *
  * Instances of this object exist embedded within ptlrpc_request; the main
@@ -1447,6 +1737,16 @@ struct ptlrpc_nrs_request {
                 * Fields for the FIFO policy
                 */
                struct nrs_fifo_req     fifo;
+               /**
+                * CRR-N request defintion
+                */
+               struct nrs_crrn_req     crr;
+               /** ORR and TRR share the same request definition */
+               struct nrs_orr_req      orr;
+               /**
+                * TBF request definition
+                */
+               struct nrs_tbf_req      tbf;
        } nr_u;
        /**
         * Externally-registering policies may want to use this to allocate
@@ -1490,34 +1790,34 @@ struct ptlrpc_hpreq_ops {
  */
 struct ptlrpc_request {
        /* Request type: one of PTL_RPC_MSG_* */
-       int rq_type;
+       int                     rq_type;
        /** Result of request processing */
-       int rq_status;
-        /**
-         * Linkage item through which this request is included into
-         * sending/delayed lists on client and into rqbd list on server
-         */
-        cfs_list_t rq_list;
-        /**
-         * Server side list of incoming unserved requests sorted by arrival
-         * time.  Traversed from time to time to notice about to expire
-         * requests and sent back "early replies" to clients to let them
-         * know server is alive and well, just very busy to service their
-         * requests in time
-         */
-        cfs_list_t rq_timed_list;
-        /** server-side history, used for debuging purposes. */
-        cfs_list_t rq_history_list;
-        /** server-side per-export list */
-        cfs_list_t rq_exp_list;
-        /** server-side hp handlers */
-        struct ptlrpc_hpreq_ops *rq_ops;
+       int                     rq_status;
+       /**
+        * Linkage item through which this request is included into
+        * sending/delayed lists on client and into rqbd list on server
+        */
+       struct list_head        rq_list;
+       /**
+        * Server side list of incoming unserved requests sorted by arrival
+        * time.  Traversed from time to time to notice about to expire
+        * requests and sent back "early replies" to clients to let them
+        * know server is alive and well, just very busy to service their
+        * requests in time
+        */
+       struct list_head        rq_timed_list;
+       /** server-side history, used for debuging purposes. */
+       struct list_head        rq_history_list;
+       /** server-side per-export list */
+       struct list_head        rq_exp_list;
+       /** server-side hp handlers */
+       struct ptlrpc_hpreq_ops *rq_ops;
 
        /** initial thread servicing this request */
-       struct ptlrpc_thread *rq_svc_thread;
+       struct ptlrpc_thread    *rq_svc_thread;
 
         /** history sequence # */
-        __u64 rq_history_seq;
+       __u64                   rq_history_seq;
        /** \addtogroup  nrs
         * @{
         */
@@ -1545,7 +1845,8 @@ struct ptlrpc_request {
                 rq_replay:1,
                 rq_no_resend:1, rq_waiting:1, rq_receiving_reply:1,
                 rq_no_delay:1, rq_net_err:1, rq_wait_ctx:1,
-                rq_early:1, rq_must_unlink:1,
+               rq_early:1,
+               rq_req_unlink:1, rq_reply_unlink:1,
                 rq_memalloc:1,      /* req originated from "kswapd" */
                 /* server-side flags */
                 rq_packed_final:1,  /* packed final reply */
@@ -1560,14 +1861,16 @@ struct ptlrpc_request {
                rq_no_retry_einprogress:1,
                /* allow the req to be sent if the import is in recovery
                 * status */
-               rq_allow_replay:1;
+               rq_allow_replay:1,
+               /* bulk request, sent to server, but uncommitted */
+               rq_unstable:1;
 
        unsigned int rq_nr_resend;
 
-        enum rq_phase rq_phase; /* one of RQ_PHASE_* */
-        enum rq_phase rq_next_phase; /* one of RQ_PHASE_* to be used next */
-        cfs_atomic_t rq_refcount;/* client-side refcount for SENT race,
-                                    server-side refcounf for multiple replies */
+       enum rq_phase rq_phase; /* one of RQ_PHASE_* */
+       enum rq_phase rq_next_phase; /* one of RQ_PHASE_* to be used next */
+       atomic_t rq_refcount;/* client-side refcount for SENT race,
+                                   server-side refcounf for multiple replies */
 
        /** Portal to which this request would be sent */
        short rq_request_portal;  /* XXX FIXME bug 249 */
@@ -1597,17 +1900,17 @@ struct ptlrpc_request {
          * there.
          * Also see \a rq_replay comment above.
          */
-        cfs_list_t rq_replay_list;
+       struct list_head        rq_replay_list;
 
-        /**
-         * security and encryption data
-         * @{ */
-        struct ptlrpc_cli_ctx   *rq_cli_ctx;     /**< client's half ctx */
-        struct ptlrpc_svc_ctx   *rq_svc_ctx;     /**< server's half ctx */
-        cfs_list_t               rq_ctx_chain;   /**< link to waited ctx */
+       /**
+        * security and encryption data
+        * @{ */
+       struct ptlrpc_cli_ctx   *rq_cli_ctx;    /**< client's half ctx */
+       struct ptlrpc_svc_ctx   *rq_svc_ctx;    /**< server's half ctx */
+       struct list_head        rq_ctx_chain;   /**< link to waited ctx */
 
-        struct sptlrpc_flavor    rq_flvr;        /**< for client & server */
-        enum lustre_sec_part     rq_sp_from;
+       struct sptlrpc_flavor   rq_flvr;        /**< for client & server */
+       enum lustre_sec_part    rq_sp_from;
 
        /* client/server security flags */
        unsigned int
@@ -1678,10 +1981,10 @@ struct ptlrpc_request {
         /** incoming request buffer */
         struct ptlrpc_request_buffer_desc *rq_rqbd;
 
-        /** client-only incoming reply */
-        lnet_handle_md_t     rq_reply_md_h;
-        cfs_waitq_t          rq_reply_waitq;
-        struct ptlrpc_cb_id  rq_reply_cbid;
+       /** client-only incoming reply */
+       lnet_handle_md_t     rq_reply_md_h;
+       wait_queue_head_t    rq_reply_waitq;
+       struct ptlrpc_cb_id  rq_reply_cbid;
 
         /** our LNet NID */
         lnet_nid_t           rq_self;
@@ -1729,9 +2032,9 @@ struct ptlrpc_request {
 
         /** Multi-rpc bits */
         /** Per-request waitq introduced by bug 21938 for recovery waiting */
-        cfs_waitq_t rq_set_waitq;
+       wait_queue_head_t rq_set_waitq;
        /** Link item for request set lists */
-       cfs_list_t  rq_set_chain;
+       struct list_head  rq_set_chain;
         /** Link back to the request set */
         struct ptlrpc_request_set *rq_set;
         /** Async completion handler, called when reply is received */
@@ -1743,7 +2046,6 @@ struct ptlrpc_request {
         struct ptlrpc_request_pool *rq_pool;
 
         struct lu_context           rq_session;
-        struct lu_context           rq_recov_session;
 
         /** request format description */
         struct req_capsule          rq_pill;
@@ -1942,16 +2244,16 @@ do {                                                                          \
  * Structure that defines a single page of a bulk transfer
  */
 struct ptlrpc_bulk_page {
-        /** Linkage to list of pages in a bulk */
-        cfs_list_t       bp_link;
-        /**
-         * Number of bytes in a page to transfer starting from \a bp_pageoffset
-         */
-        int              bp_buflen;
-        /** offset within a page */
-        int              bp_pageoffset;
-        /** The page itself */
-        struct page     *bp_page;
+       /** Linkage to list of pages in a bulk */
+       struct list_head bp_link;
+       /**
+        * Number of bytes in a page to transfer starting from \a bp_pageoffset
+        */
+       int              bp_buflen;
+       /** offset within a page */
+       int              bp_pageoffset;
+       /** The page itself */
+       struct page     *bp_page;
 };
 
 #define BULK_GET_SOURCE   0
@@ -1988,7 +2290,7 @@ struct ptlrpc_bulk_desc {
        struct obd_import *bd_import;
        /** Back pointer to the request */
        struct ptlrpc_request *bd_req;
-       cfs_waitq_t            bd_waitq;        /* server side only WQ */
+       wait_queue_head_t      bd_waitq;        /* server side only WQ */
        int                    bd_iov_count;    /* # entries in bd_iov */
        int                    bd_max_iov;      /* allocated size of bd_iov */
        int                    bd_nob;          /* # bytes covered */
@@ -2032,7 +2334,7 @@ struct ptlrpc_thread {
         /**
          * List of active threads in svc->srv_threads
          */
-        cfs_list_t t_link;
+       struct list_head t_link;
         /**
          * thread-private data (preallocated memory)
          */
@@ -2054,7 +2356,7 @@ struct ptlrpc_thread {
          * the svc this thread belonged to b=18582
          */
        struct ptlrpc_service_part      *t_svcpt;
-       cfs_waitq_t                     t_ctl_waitq;
+       wait_queue_head_t               t_ctl_waitq;
        struct lu_env                   *t_env;
        char                            t_name[PTLRPC_THR_NAME_LEN];
 };
@@ -2127,23 +2429,23 @@ static inline int thread_test_and_clear_flags(struct ptlrpc_thread *thread,
  * More than one request can fit into the buffer.
  */
 struct ptlrpc_request_buffer_desc {
-        /** Link item for rqbds on a service */
-        cfs_list_t             rqbd_list;
-        /** History of requests for this buffer */
-        cfs_list_t             rqbd_reqs;
-        /** Back pointer to service for which this buffer is registered */
-       struct ptlrpc_service_part *rqbd_svcpt;
-        /** LNet descriptor */
-        lnet_handle_md_t       rqbd_md_h;
-        int                    rqbd_refcount;
-        /** The buffer itself */
-        char                  *rqbd_buffer;
-        struct ptlrpc_cb_id    rqbd_cbid;
-        /**
-         * This "embedded" request structure is only used for the
-         * last request to fit into the buffer
-         */
-        struct ptlrpc_request  rqbd_req;
+       /** Link item for rqbds on a service */
+       struct list_head                rqbd_list;
+       /** History of requests for this buffer */
+       struct list_head                rqbd_reqs;
+       /** Back pointer to service for which this buffer is registered */
+       struct ptlrpc_service_part      *rqbd_svcpt;
+       /** LNet descriptor */
+       lnet_handle_md_t                rqbd_md_h;
+       int                             rqbd_refcount;
+       /** The buffer itself */
+       char                            *rqbd_buffer;
+       struct ptlrpc_cb_id             rqbd_cbid;
+       /**
+        * This "embedded" request structure is only used for the
+        * last request to fit into the buffer
+        */
+       struct ptlrpc_request           rqbd_req;
 };
 
 typedef int  (*svc_handler_t)(struct ptlrpc_request *req);
@@ -2194,9 +2496,9 @@ struct ptlrpc_service_ops {
 struct ptlrpc_service {
        /** serialize /proc operations */
        spinlock_t                      srv_lock;
-        /** most often accessed fields */
-        /** chain thru all services */
-        cfs_list_t                      srv_list;
+       /** most often accessed fields */
+       /** chain thru all services */
+       struct list_head                srv_list;
        /** service operations table */
        struct ptlrpc_service_ops       srv_ops;
         /** only statically allocated strings here; we don't clean them */
@@ -2204,7 +2506,7 @@ struct ptlrpc_service {
         /** only statically allocated strings here; we don't clean them */
         char                           *srv_thread_name;
         /** service thread list */
-        cfs_list_t                      srv_threads;
+       struct list_head                srv_threads;
        /** threads # should be created for each partition on initializing */
        int                             srv_nthrs_cpt_init;
        /** limit of threads number for each partition */
@@ -2285,7 +2587,7 @@ struct ptlrpc_service_part {
        /** # running threads */
        int                             scp_nthrs_running;
        /** service threads list */
-       cfs_list_t                      scp_threads;
+       struct list_head                scp_threads;
 
        /**
         * serialize the following fields, used for protecting
@@ -2302,23 +2604,23 @@ struct ptlrpc_service_part {
        /** # incoming reqs */
        int                             scp_nreqs_incoming;
        /** request buffers to be reposted */
-       cfs_list_t                      scp_rqbd_idle;
+       struct list_head                scp_rqbd_idle;
        /** req buffers receiving */
-       cfs_list_t                      scp_rqbd_posted;
+       struct list_head                scp_rqbd_posted;
        /** incoming reqs */
-       cfs_list_t                      scp_req_incoming;
+       struct list_head                scp_req_incoming;
        /** timeout before re-posting reqs, in tick */
        cfs_duration_t                  scp_rqbd_timeout;
        /**
         * all threads sleep on this. This wait-queue is signalled when new
         * incoming request arrives and when difficult reply has to be handled.
         */
-       cfs_waitq_t                     scp_waitq;
+       wait_queue_head_t               scp_waitq;
 
        /** request history */
-       cfs_list_t                      scp_hist_reqs;
+       struct list_head                scp_hist_reqs;
        /** request buffer history */
-       cfs_list_t                      scp_hist_rqbds;
+       struct list_head                scp_hist_rqbds;
        /** # request buffers in history */
        int                             scp_hist_nrqbds;
        /** sequence number for request */
@@ -2357,7 +2659,7 @@ struct ptlrpc_service_part {
        /** reqs waiting for replies */
        struct ptlrpc_at_array          scp_at_array;
        /** early reply timer */
-       cfs_timer_t                     scp_at_timer;
+       struct timer_list               scp_at_timer;
        /** debug */
        cfs_time_t                      scp_at_checktime;
        /** check early replies */
@@ -2370,17 +2672,17 @@ struct ptlrpc_service_part {
         */
        spinlock_t                      scp_rep_lock __cfs_cacheline_aligned;
        /** all the active replies */
-       cfs_list_t                      scp_rep_active;
+       struct list_head                scp_rep_active;
 #ifndef __KERNEL__
        /** replies waiting for service */
-       cfs_list_t                      scp_rep_queue;
+       struct list_head                scp_rep_queue;
 #endif
        /** List of free reply_states */
-       cfs_list_t                      scp_rep_idle;
+       struct list_head                scp_rep_idle;
        /** waitq to run, when adding stuff to srv_free_rs_list */
-       cfs_waitq_t                     scp_rep_waitq;
+       wait_queue_head_t               scp_rep_waitq;
        /** # 'difficult' replies */
-       cfs_atomic_t                    scp_nreps_difficult;
+       atomic_t                        scp_nreps_difficult;
 };
 
 #define ptlrpc_service_for_each_part(part, i, svc)                     \
@@ -2414,7 +2716,7 @@ struct ptlrpcd_ctl {
          */
         struct ptlrpc_request_set  *pc_set;
         /**
-         * Thread name used in cfs_daemonize()
+        * Thread name used in kthread_run()
          */
         char                        pc_name[16];
         /**
@@ -2621,6 +2923,8 @@ int ptlrpc_register_rqbd(struct ptlrpc_request_buffer_desc *rqbd);
  * request queues, request management, etc.
  * @{
  */
+void ptlrpc_request_committed(struct ptlrpc_request *req, int force);
+
 void ptlrpc_init_client(int req_portal, int rep_portal, char *name,
                         struct ptlrpc_client *);
 void ptlrpc_cleanup_client(struct obd_import *imp);
@@ -2695,16 +2999,16 @@ static inline void ptlrpc_free_bulk_nopin(struct ptlrpc_bulk_desc *bulk)
        __ptlrpc_free_bulk(bulk, 0);
 }
 void __ptlrpc_prep_bulk_page(struct ptlrpc_bulk_desc *desc,
-                            cfs_page_t *page, int pageoffset, int len, int);
+                            struct page *page, int pageoffset, int len, int);
 static inline void ptlrpc_prep_bulk_page_pin(struct ptlrpc_bulk_desc *desc,
-                                            cfs_page_t *page, int pageoffset,
+                                            struct page *page, int pageoffset,
                                             int len)
 {
        __ptlrpc_prep_bulk_page(desc, page, pageoffset, len, 1);
 }
 
 static inline void ptlrpc_prep_bulk_page_nopin(struct ptlrpc_bulk_desc *desc,
-                                              cfs_page_t *page, int pageoffset,
+                                              struct page *page, int pageoffset,
                                               int len)
 {
        __ptlrpc_prep_bulk_page(desc, page, pageoffset, len, 0);
@@ -2808,6 +3112,8 @@ int liblustre_check_services(void *arg);
 void ptlrpc_daemonize(char *name);
 int ptlrpc_service_health_check(struct ptlrpc_service *);
 void ptlrpc_server_drop_request(struct ptlrpc_request *req);
+void ptlrpc_request_change_export(struct ptlrpc_request *req,
+                                 struct obd_export *export);
 
 #ifdef __KERNEL__
 int ptlrpc_hr_init(void);
@@ -2935,30 +3241,62 @@ lustre_shrink_reply(struct ptlrpc_request *req, int segment,
         req->rq_replen = lustre_shrink_msg(req->rq_repmsg, segment,
                                            newlen, move_data);
 }
+
+#ifdef LUSTRE_TRANSLATE_ERRNOS
+
+static inline int ptlrpc_status_hton(int h)
+{
+       /*
+        * Positive errnos must be network errnos, such as LUSTRE_EDEADLK,
+        * ELDLM_LOCK_ABORTED, etc.
+        */
+       if (h < 0)
+               return -lustre_errno_hton(-h);
+       else
+               return h;
+}
+
+static inline int ptlrpc_status_ntoh(int n)
+{
+       /*
+        * See the comment in ptlrpc_status_hton().
+        */
+       if (n < 0)
+               return -lustre_errno_ntoh(-n);
+       else
+               return n;
+}
+
+#else
+
+#define ptlrpc_status_hton(h) (h)
+#define ptlrpc_status_ntoh(n) (n)
+
+#endif
 /** @} */
 
 /** Change request phase of \a req to \a new_phase */
 static inline void
 ptlrpc_rqphase_move(struct ptlrpc_request *req, enum rq_phase new_phase)
 {
-        if (req->rq_phase == new_phase)
-                return;
+       if (req->rq_phase == new_phase)
+               return;
 
-        if (new_phase == RQ_PHASE_UNREGISTERING) {
-                req->rq_next_phase = req->rq_phase;
-                if (req->rq_import)
-                        cfs_atomic_inc(&req->rq_import->imp_unregistering);
-        }
+       if (new_phase == RQ_PHASE_UNREGISTERING) {
+               req->rq_next_phase = req->rq_phase;
+               if (req->rq_import)
+                       atomic_inc(&req->rq_import->imp_unregistering);
+       }
 
-        if (req->rq_phase == RQ_PHASE_UNREGISTERING) {
-                if (req->rq_import)
-                        cfs_atomic_dec(&req->rq_import->imp_unregistering);
-        }
+       if (req->rq_phase == RQ_PHASE_UNREGISTERING) {
+               if (req->rq_import)
+                       atomic_dec(&req->rq_import->imp_unregistering);
+       }
 
-        DEBUG_REQ(D_INFO, req, "move req \"%s\" -> \"%s\"",
-                  ptlrpc_rqphase2str(req), ptlrpc_phase2str(new_phase));
+       DEBUG_REQ(D_INFO, req, "move req \"%s\" -> \"%s\"",
+                 ptlrpc_rqphase2str(req), ptlrpc_phase2str(new_phase));
 
-        req->rq_phase = new_phase;
+       req->rq_phase = new_phase;
 }
 
 /**
@@ -3006,7 +3344,8 @@ ptlrpc_client_recv_or_unlink(struct ptlrpc_request *req)
                spin_unlock(&req->rq_lock);
                return 1;
        }
-       rc = req->rq_receiving_reply || req->rq_must_unlink;
+       rc = req->rq_receiving_reply ;
+       rc = rc || req->rq_req_unlink || req->rq_reply_unlink;
        spin_unlock(&req->rq_lock);
        return rc;
 }
@@ -3014,25 +3353,25 @@ ptlrpc_client_recv_or_unlink(struct ptlrpc_request *req)
 static inline void
 ptlrpc_client_wake_req(struct ptlrpc_request *req)
 {
-        if (req->rq_set == NULL)
-                cfs_waitq_signal(&req->rq_reply_waitq);
-        else
-                cfs_waitq_signal(&req->rq_set->set_waitq);
+       if (req->rq_set == NULL)
+               wake_up(&req->rq_reply_waitq);
+       else
+               wake_up(&req->rq_set->set_waitq);
 }
 
 static inline void
 ptlrpc_rs_addref(struct ptlrpc_reply_state *rs)
 {
-        LASSERT(cfs_atomic_read(&rs->rs_refcount) > 0);
-        cfs_atomic_inc(&rs->rs_refcount);
+       LASSERT(atomic_read(&rs->rs_refcount) > 0);
+       atomic_inc(&rs->rs_refcount);
 }
 
 static inline void
 ptlrpc_rs_decref(struct ptlrpc_reply_state *rs)
 {
-        LASSERT(cfs_atomic_read(&rs->rs_refcount) > 0);
-        if (cfs_atomic_dec_and_test(&rs->rs_refcount))
-                lustre_free_reply_state(rs);
+       LASSERT(atomic_read(&rs->rs_refcount) > 0);
+       if (atomic_dec_and_test(&rs->rs_refcount))
+               lustre_free_reply_state(rs);
 }
 
 /* Should only be called once per req */
@@ -3137,13 +3476,12 @@ typedef int (*timeout_cb_t)(struct timeout_item *, void *);
 int ptlrpc_pinger_add_import(struct obd_import *imp);
 int ptlrpc_pinger_del_import(struct obd_import *imp);
 int ptlrpc_add_timeout_client(int time, enum timeout_event event,
-                              timeout_cb_t cb, void *data,
-                              cfs_list_t *obd_list);
-int ptlrpc_del_timeout_client(cfs_list_t *obd_list,
+                             timeout_cb_t cb, void *data,
+                             struct list_head *obd_list);
+int ptlrpc_del_timeout_client(struct list_head *obd_list,
                               enum timeout_event event);
 struct ptlrpc_request * ptlrpc_prep_ping(struct obd_import *imp);
 int ptlrpc_obd_ping(struct obd_device *obd);
-cfs_time_t ptlrpc_suspend_wakeup_time(void);
 #ifdef __KERNEL__
 void ping_evictor_start(void);
 void ping_evictor_stop(void);
@@ -3151,7 +3489,8 @@ void ping_evictor_stop(void);
 #define ping_evictor_start()    do {} while (0)
 #define ping_evictor_stop()     do {} while (0)
 #endif
-int ptlrpc_check_and_wait_suspend(struct ptlrpc_request *req);
+void ptlrpc_pinger_ir_up(void);
+void ptlrpc_pinger_ir_down(void);
 /** @} */
 int ptlrpc_pinger_suppress_pings(void);
 
@@ -3220,7 +3559,6 @@ int llog_origin_handle_prev_block(struct ptlrpc_request *req);
 int llog_origin_handle_next_block(struct ptlrpc_request *req);
 int llog_origin_handle_read_header(struct ptlrpc_request *req);
 int llog_origin_handle_close(struct ptlrpc_request *req);
-int llog_origin_handle_cancel(struct ptlrpc_request *req);
 
 /* ptlrpc/llog_client.c */
 extern struct llog_operations llog_client_ops;