Whamcloud - gitweb
LU-17744 ldiskfs: mballoc stats fixes
[fs/lustre-release.git] / lnet / klnds / o2iblnd / o2iblnd_modparams.c
index 57a025e..ffb68e5 100644 (file)
@@ -27,7 +27,6 @@
  */
 /*
  * This file is part of Lustre, http://www.lustre.org/
- * Lustre is a trademark of Sun Microsystems, Inc.
  *
  * lnet/klnds/o2iblnd/o2iblnd_modparams.c
  *
@@ -92,7 +91,7 @@ MODULE_PARM_DESC(ipif_name, "IPoIB interface name");
 
 static int retry_count = 5;
 module_param(retry_count, int, 0644);
-MODULE_PARM_DESC(retry_count, "Retransmissions when no ACK received");
+MODULE_PARM_DESC(retry_count, "Number of times to retry connection operations");
 
 static int rnr_retry_count = 6;
 module_param(rnr_retry_count, int, 0644);
@@ -140,14 +139,13 @@ MODULE_PARM_DESC(use_fastreg_gaps, "Enable discontiguous fastreg fragment suppor
  *  4. Look at the comments in kiblnd_fmr_map_tx() for an explanation of
  *     the behavior when transmit with GAPS verses contiguous.
  */
-#ifdef HAVE_IB_GET_DMA_MR
-#define IBLND_DEFAULT_MAP_ON_DEMAND 0
+
+#ifdef HAVE_OFED_IB_GET_DMA_MR
 #define MOD_STR "map on demand"
 #else
-#define IBLND_DEFAULT_MAP_ON_DEMAND 1
 #define MOD_STR "map on demand (obsolete)"
 #endif
-static int map_on_demand = IBLND_DEFAULT_MAP_ON_DEMAND;
+static int map_on_demand = 1;
 module_param(map_on_demand, int, 0444);
 MODULE_PARM_DESC(map_on_demand, MOD_STR);
 
@@ -186,6 +184,22 @@ static unsigned int wrq_sge = 2;
 module_param(wrq_sge, uint, 0444);
 MODULE_PARM_DESC(wrq_sge, "# scatter/gather element per work request");
 
+static int tos = -1;
+static int param_set_tos(const char *val, cfs_kernel_param_arg_t *kp);
+#ifdef HAVE_KERNEL_PARAM_OPS
+static const struct kernel_param_ops param_ops_tos = {
+       .set = param_set_tos,
+       .get = param_get_int,
+};
+
+#define param_check_tos(name, p) \
+       __param_check(name, p, int)
+module_param(tos, tos, 0444);
+#else
+module_param_call(tos, param_set_tos, param_get_int, &tos, 0444);
+#endif
+MODULE_PARM_DESC(tos, "Set the type of service (=-1 to disable)");
+
 struct kib_tunables kiblnd_tunables = {
         .kib_dev_failover           = &dev_failover,
         .kib_service                = &service,
@@ -203,7 +217,26 @@ struct kib_tunables kiblnd_tunables = {
        .kib_use_fastreg_gaps       = &use_fastreg_gaps,
 };
 
-static struct lnet_ioctl_config_o2iblnd_tunables default_tunables;
+struct lnet_ioctl_config_o2iblnd_tunables kib_default_tunables;
+
+static int param_set_tos(const char *val, cfs_kernel_param_arg_t *kp)
+{
+       int rc, t;
+
+       if (!val)
+               return -EINVAL;
+
+       rc = kstrtoint(val, 0, &t);
+       if (rc)
+               return rc;
+
+       if (t < -1 || t > 0xff)
+               return -ERANGE;
+
+       *((int *)kp->arg) = t;
+
+       return 0;
+}
 
 /* # messages/RDMAs in-flight */
 int
@@ -229,7 +262,7 @@ kiblnd_tunables_setup(struct lnet_ni *ni)
         */
        if (!ni->ni_lnd_tunables_set)
                memcpy(&ni->ni_lnd_tunables.lnd_tun_u.lnd_o2ib,
-                      &default_tunables, sizeof(*tunables));
+                      &kib_default_tunables, sizeof(*tunables));
 
        tunables = &ni->ni_lnd_tunables.lnd_tun_u.lnd_o2ib;
 
@@ -269,7 +302,10 @@ kiblnd_tunables_setup(struct lnet_ni *ni)
                net_tunables->lct_peer_tx_credits =
                        net_tunables->lct_max_tx_credits;
 
-#ifndef HAVE_IB_GET_DMA_MR
+       if (tunables->lnd_map_on_demand == UINT_MAX)
+               tunables->lnd_map_on_demand = map_on_demand;
+
+#ifndef HAVE_OFED_IB_GET_DMA_MR
        /*
         * For kernels which do not support global memory regions, always
         * enable map_on_demand
@@ -311,10 +347,13 @@ kiblnd_tunables_setup(struct lnet_ni *ni)
                tunables->lnd_fmr_cache = fmr_cache;
        if (!tunables->lnd_ntx)
                tunables->lnd_ntx = ntx;
-       if (!tunables->lnd_conns_per_peer) {
+       if (!tunables->lnd_conns_per_peer)
                tunables->lnd_conns_per_peer = (conns_per_peer) ?
                        conns_per_peer : 1;
-       }
+       if (tunables->lnd_tos < 0)
+               tunables->lnd_tos = tos;
+
+       tunables->lnd_timeout = kiblnd_timeout();
 
        return 0;
 }
@@ -322,14 +361,15 @@ kiblnd_tunables_setup(struct lnet_ni *ni)
 int
 kiblnd_tunables_init(void)
 {
-       default_tunables.lnd_version = CURRENT_LND_VERSION;
-       default_tunables.lnd_peercredits_hiw = peer_credits_hiw;
-       default_tunables.lnd_map_on_demand = map_on_demand;
-       default_tunables.lnd_concurrent_sends = concurrent_sends;
-       default_tunables.lnd_fmr_pool_size = fmr_pool_size;
-       default_tunables.lnd_fmr_flush_trigger = fmr_flush_trigger;
-       default_tunables.lnd_fmr_cache = fmr_cache;
-       default_tunables.lnd_ntx = ntx;
-       default_tunables.lnd_conns_per_peer = conns_per_peer;
+       kib_default_tunables.lnd_version = CURRENT_LND_VERSION;
+       kib_default_tunables.lnd_peercredits_hiw = peer_credits_hiw;
+       kib_default_tunables.lnd_map_on_demand = map_on_demand;
+       kib_default_tunables.lnd_concurrent_sends = concurrent_sends;
+       kib_default_tunables.lnd_fmr_pool_size = fmr_pool_size;
+       kib_default_tunables.lnd_fmr_flush_trigger = fmr_flush_trigger;
+       kib_default_tunables.lnd_fmr_cache = fmr_cache;
+       kib_default_tunables.lnd_ntx = ntx;
+       kib_default_tunables.lnd_conns_per_peer = conns_per_peer;
+       kib_default_tunables.lnd_tos = tos;
        return 0;
 }