Whamcloud - gitweb
LU-11299 lnet: lnet_add/del_route()
[fs/lustre-release.git] / lnet / lnet / router.c
index 42df3a2..4aefdf1 100644 (file)
@@ -1,10 +1,9 @@
-/* -*- mode: c; c-basic-offset: 8; indent-tabs-mode: nil; -*-
- * vim:expandtab:shiftwidth=8:tabstop=8:
+/*
+ * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
  *
- * Copyright  2008 Sun Microsystems, Inc. All rights reserved
+ * Copyright (c) 2011, 2017, Intel Corporation.
  *
- *   This file is part of Portals
- *   http://sourceforge.net/projects/sandiaportals/
+ *   This file is part of Lustre, https://wiki.whamcloud.com/
  *
  *   Portals is free software; you can redistribute it and/or
  *   modify it under the terms of version 2 of the GNU General Public
  */
 
 #define DEBUG_SUBSYSTEM S_LNET
+
+#include <linux/random.h>
 #include <lnet/lib-lnet.h>
 
-#if defined(__KERNEL__) && defined(LNET_ROUTER)
+#define LNET_NRB_TINY_MIN      512     /* min value for each CPT */
+#define LNET_NRB_TINY          (LNET_NRB_TINY_MIN * 4)
+#define LNET_NRB_SMALL_MIN     4096    /* min value for each CPT */
+#define LNET_NRB_SMALL         (LNET_NRB_SMALL_MIN * 4)
+#define LNET_NRB_SMALL_PAGES   1
+#define LNET_NRB_LARGE_MIN     256     /* min value for each CPT */
+#define LNET_NRB_LARGE         (LNET_NRB_LARGE_MIN * 4)
+#define LNET_NRB_LARGE_PAGES   ((LNET_MTU + PAGE_SIZE - 1) >> \
+                                 PAGE_SHIFT)
 
 static char *forwarding = "";
-CFS_MODULE_PARM(forwarding, "s", charp, 0444,
-                "Explicitly enable/disable forwarding between networks");
-
-static int tiny_router_buffers = 1024;
-CFS_MODULE_PARM(tiny_router_buffers, "i", int, 0444,
-                "# of 0 payload messages to buffer in the router");
-static int small_router_buffers = 8192;
-CFS_MODULE_PARM(small_router_buffers, "i", int, 0444,
-                "# of small (1 page) messages to buffer in the router");
-static int large_router_buffers = 512;
-CFS_MODULE_PARM(large_router_buffers, "i", int, 0444,
-                "# of large messages to buffer in the router");
-static int peer_buffer_credits = 0;
-CFS_MODULE_PARM(peer_buffer_credits, "i", int, 0444,
-                "# router buffer credits per peer");
+module_param(forwarding, charp, 0444);
+MODULE_PARM_DESC(forwarding, "Explicitly enable/disable forwarding between networks");
+
+static int tiny_router_buffers;
+module_param(tiny_router_buffers, int, 0444);
+MODULE_PARM_DESC(tiny_router_buffers, "# of 0 payload messages to buffer in the router");
+static int small_router_buffers;
+module_param(small_router_buffers, int, 0444);
+MODULE_PARM_DESC(small_router_buffers, "# of small (1 page) messages to buffer in the router");
+static int large_router_buffers;
+module_param(large_router_buffers, int, 0444);
+MODULE_PARM_DESC(large_router_buffers, "# of large messages to buffer in the router");
+static int peer_buffer_credits;
+module_param(peer_buffer_credits, int, 0444);
+MODULE_PARM_DESC(peer_buffer_credits, "# router buffer credits per peer");
 
 static int auto_down = 1;
-CFS_MODULE_PARM(auto_down, "i", int, 0444,
-                "Automatically mark peers down on comms error");
+module_param(auto_down, int, 0444);
+MODULE_PARM_DESC(auto_down, "Automatically mark peers down on comms error");
+
+int
+lnet_peer_buffer_credits(struct lnet_net *net)
+{
+       /* NI option overrides LNet default */
+       if (net->net_tunables.lct_peer_rtr_credits > 0)
+               return net->net_tunables.lct_peer_rtr_credits;
+       if (peer_buffer_credits > 0)
+               return peer_buffer_credits;
+
+       /* As an approximation, allow this peer the same number of router
+        * buffers as it is allowed outstanding sends */
+       return net->net_tunables.lct_peer_tx_credits;
+}
 
-static int check_routers_before_use = 0;
-CFS_MODULE_PARM(check_routers_before_use, "i", int, 0444,
-                "Assume routers are down and ping them before use");
+static int check_routers_before_use;
+module_param(check_routers_before_use, int, 0444);
+MODULE_PARM_DESC(check_routers_before_use, "Assume routers are down and ping them before use");
 
-static int dead_router_check_interval = 0;
-CFS_MODULE_PARM(dead_router_check_interval, "i", int, 0444,
-                "Seconds between dead router health checks (<= 0 to disable)");
+int avoid_asym_router_failure = 1;
+module_param(avoid_asym_router_failure, int, 0644);
+MODULE_PARM_DESC(avoid_asym_router_failure, "Avoid asymmetrical router failures (0 to disable)");
 
-static int live_router_check_interval = 0;
-CFS_MODULE_PARM(live_router_check_interval, "i", int, 0444,
-                "Seconds between live router health checks (<= 0 to disable)");
+static int dead_router_check_interval = 60;
+module_param(dead_router_check_interval, int, 0644);
+MODULE_PARM_DESC(dead_router_check_interval, "Seconds between dead router health checks (<= 0 to disable)");
+
+static int live_router_check_interval = 60;
+module_param(live_router_check_interval, int, 0644);
+MODULE_PARM_DESC(live_router_check_interval, "Seconds between live router health checks (<= 0 to disable)");
 
 static int router_ping_timeout = 50;
-CFS_MODULE_PARM(router_ping_timeout, "i", int, 0444,
-                "Seconds to wait for the reply to a router health query");
+module_param(router_ping_timeout, int, 0644);
+MODULE_PARM_DESC(router_ping_timeout, "Seconds to wait for the reply to a router health query");
 
 int
 lnet_peers_start_down(void)
 {
-        return check_routers_before_use;
+       return check_routers_before_use;
 }
 
-int
-lnet_peer_buffer_credits(lnet_ni_t *ni)
+void
+lnet_notify_locked(struct lnet_peer_ni *lp, int notifylnd, int alive,
+                  time64_t when)
 {
-        /* NI option overrides LNet default */
-        if (ni->ni_peerrtrcredits > 0)
-                return ni->ni_peerrtrcredits;
-        if (peer_buffer_credits > 0)
-                return peer_buffer_credits;
-
-        /* As an approximation, allow this peer the same number of router
-         * buffers as it is allowed outstanding sends */
-        return ni->ni_peertxcredits;
+       if (lp->lpni_timestamp > when) { /* out of date information */
+               CDEBUG(D_NET, "Out of date\n");
+               return;
+       }
+
+       /*
+        * This function can be called with different cpt locks being
+        * held. lpni_alive_count modification needs to be properly protected.
+        * Significant reads to lpni_alive_count are also protected with
+        * the same lock
+        */
+       spin_lock(&lp->lpni_lock);
+
+       lp->lpni_timestamp = when; /* update timestamp */
+
+       if (lp->lpni_alive_count != 0 &&          /* got old news */
+           (!lp->lpni_alive) == (!alive)) {      /* new date for old news */
+               spin_unlock(&lp->lpni_lock);
+               CDEBUG(D_NET, "Old news\n");
+               return;
+       }
+
+       /* Flag that notification is outstanding */
+
+       lp->lpni_alive_count++;
+       lp->lpni_alive = (alive) ? 1 : 0;
+       lp->lpni_notify = 1;
+       lp->lpni_notifylnd = notifylnd;
+       if (lp->lpni_alive)
+               lp->lpni_ping_feats = LNET_PING_FEAT_INVAL; /* reset */
+
+       spin_unlock(&lp->lpni_lock);
+
+       CDEBUG(D_NET, "set %s %d\n", libcfs_nid2str(lp->lpni_nid), alive);
 }
 
-void
-lnet_notify_locked(lnet_peer_t *lp, int notifylnd, int alive, time_t when)
+/*
+ * This function will always be called with lp->lpni_cpt lock held.
+ */
+static void
+lnet_ni_notify_locked(struct lnet_ni *ni, struct lnet_peer_ni *lp)
 {
-        if (when < lp->lp_timestamp) {          /* out of date information */
-                CDEBUG(D_NET, "Out of date\n");
-                return;
-        }
-
-        lp->lp_timestamp = when;                /* update timestamp */
-        lp->lp_ping_deadline = 0;               /* disable ping timeout */
+       int alive;
+       int notifylnd;
 
-        if (lp->lp_alive_count != 0 &&          /* got old news */
-            (!lp->lp_alive) == (!alive)) {      /* new date for old news */
-                CDEBUG(D_NET, "Old news\n");
-                return;
-        }
+       /* Notify only in 1 thread at any time to ensure ordered notification.
+        * NB individual events can be missed; the only guarantee is that you
+        * always get the most recent news */
 
-        /* Flag that notification is outstanding */
+       spin_lock(&lp->lpni_lock);
 
-        lp->lp_alive_count++;
-        lp->lp_alive = !(!alive);               /* 1 bit! */
-        lp->lp_notify = 1;
-        lp->lp_notifylnd |= notifylnd;
+       if (lp->lpni_notifying || ni == NULL) {
+               spin_unlock(&lp->lpni_lock);
+               return;
+       }
 
-        CDEBUG(D_NET, "set %s %d\n", libcfs_nid2str(lp->lp_nid), alive);
-}
+       lp->lpni_notifying = 1;
 
-void
-lnet_do_notify (lnet_peer_t *lp)
-{
-        lnet_ni_t *ni = lp->lp_ni;
-        int        alive;
-        int        notifylnd;
+       /*
+        * lp->lpni_notify needs to be protected because it can be set in
+        * lnet_notify_locked().
+        */
+       while (lp->lpni_notify) {
+               alive     = lp->lpni_alive;
+               notifylnd = lp->lpni_notifylnd;
 
-        LNET_LOCK();
+               lp->lpni_notifylnd = 0;
+               lp->lpni_notify    = 0;
 
-        /* Notify only in 1 thread at any time to ensure ordered notification.
-         * NB individual events can be missed; the only guarantee is that you
-         * always get the most recent news */
+               if (notifylnd && ni->ni_net->net_lnd->lnd_notify != NULL) {
+                       spin_unlock(&lp->lpni_lock);
+                       lnet_net_unlock(lp->lpni_cpt);
 
-        if (lp->lp_notifying) {
-                LNET_UNLOCK();
-                return;
-        }
+                       /* A new notification could happen now; I'll handle it
+                        * when control returns to me */
 
-        lp->lp_notifying = 1;
+                       (ni->ni_net->net_lnd->lnd_notify)(ni, lp->lpni_nid,
+                                                         alive);
 
-        while (lp->lp_notify) {
-                alive     = lp->lp_alive;
-                notifylnd = lp->lp_notifylnd;
+                       lnet_net_lock(lp->lpni_cpt);
+                       spin_lock(&lp->lpni_lock);
+               }
+       }
 
-                lp->lp_notifylnd = 0;
-                lp->lp_notify    = 0;
+       lp->lpni_notifying = 0;
+       spin_unlock(&lp->lpni_lock);
+}
 
-                if (notifylnd && ni->ni_lnd->lnd_notify != NULL) {
-                        LNET_UNLOCK();
+static void
+lnet_rtr_addref_locked(struct lnet_peer *lp)
+{
+       LASSERT(lp->lp_rtr_refcount >= 0);
+
+       /* lnet_net_lock must be exclusively locked */
+       lp->lp_rtr_refcount++;
+       if (lp->lp_rtr_refcount == 1) {
+               list_add_tail(&lp->lp_rtr_list, &the_lnet.ln_routers);
+               /* addref for the_lnet.ln_routers */
+               lnet_peer_addref_locked(lp);
+               the_lnet.ln_routers_version++;
+       }
+}
 
-                        /* A new notification could happen now; I'll handle it
-                         * when control returns to me */
+static void
+lnet_rtr_decref_locked(struct lnet_peer *lp)
+{
+       LASSERT(atomic_read(&lp->lp_refcount) > 0);
+       LASSERT(lp->lp_rtr_refcount > 0);
+
+       /* lnet_net_lock must be exclusively locked */
+       lp->lp_rtr_refcount--;
+       if (lp->lp_rtr_refcount == 0) {
+               LASSERT(list_empty(&lp->lp_routes));
+
+               list_del(&lp->lp_rtr_list);
+               /* decref for the_lnet.ln_routers */
+               lnet_peer_decref_locked(lp);
+               the_lnet.ln_routers_version++;
+       }
+}
 
-                        (ni->ni_lnd->lnd_notify)(ni, lp->lp_nid, alive);
+struct lnet_remotenet *
+lnet_find_rnet_locked(__u32 net)
+{
+       struct lnet_remotenet *rnet;
+       struct list_head *tmp;
+       struct list_head *rn_list;
 
-                        LNET_LOCK();
-                }
-        }
+       LASSERT(the_lnet.ln_state == LNET_STATE_RUNNING);
 
-        lp->lp_notifying = 0;
+       rn_list = lnet_net2rnethash(net);
+       list_for_each(tmp, rn_list) {
+               rnet = list_entry(tmp, struct lnet_remotenet, lrn_list);
 
-        LNET_UNLOCK();
+               if (rnet->lrn_net == net)
+                       return rnet;
+       }
+       return NULL;
 }
 
-int
-lnet_notify (lnet_ni_t *ni, lnet_nid_t nid, int alive, time_t when)
+static void lnet_shuffle_seed(void)
 {
-        lnet_peer_t         *lp = NULL;
-        time_t               now = cfs_time_current_sec();
-
-        LASSERT (!in_interrupt ());
-
-        CDEBUG (D_NET, "%s notifying %s: %s\n",
-                (ni == NULL) ? "userspace" : libcfs_nid2str(ni->ni_nid),
-                libcfs_nid2str(nid),
-                alive ? "up" : "down");
-
-        if (ni != NULL &&
-            LNET_NIDNET(ni->ni_nid) != LNET_NIDNET(nid)) {
-                CWARN ("Ignoring notification of %s %s by %s (different net)\n",
-                        libcfs_nid2str(nid), alive ? "birth" : "death",
-                        libcfs_nid2str(ni->ni_nid));
-                return -EINVAL;
-        }
-
-        /* can't do predictions... */
-        if (when > now) {
-                CWARN ("Ignoring prediction from %s of %s %s "
-                       "%ld seconds in the future\n",
-                       (ni == NULL) ? "userspace" : libcfs_nid2str(ni->ni_nid),
-                       libcfs_nid2str(nid), alive ? "up" : "down",
-                       when - now);
-                return -EINVAL;
-        }
-
-        if (ni != NULL && !alive &&             /* LND telling me she's down */
-            !auto_down) {                       /* auto-down disabled */
-                CDEBUG(D_NET, "Auto-down disabled\n");
-                return 0;
-        }
-
-        LNET_LOCK();
-
-        lp = lnet_find_peer_locked(nid);
-        if (lp == NULL) {
-                /* nid not found */
-                LNET_UNLOCK();
-                CDEBUG(D_NET, "%s not found\n", libcfs_nid2str(nid));
-                return 0;
-        }
-
-        /* We can't fully trust LND on reporting exact peer last_alive
-         * if he notifies us about dead peer. For example ksocklnd can
-         * call us with when == _time_when_the_node_was_booted_ if
-         * no connections were successfully established */
-        if (ni != NULL && !alive && when < lp->lp_last_alive)
-                when = lp->lp_last_alive;
-
-        lnet_notify_locked(lp, ni == NULL, alive, when);
-
-        LNET_UNLOCK();
-
-        lnet_do_notify(lp);
-
-        LNET_LOCK();
-
-        lnet_peer_decref_locked(lp);
-
-        LNET_UNLOCK();
-        return 0;
-}
-EXPORT_SYMBOL(lnet_notify);
+       static int seeded;
+       struct lnet_ni *ni = NULL;
 
-#else
+       if (seeded)
+               return;
 
-int
-lnet_notify (lnet_ni_t *ni, lnet_nid_t nid, int alive, time_t when)
-{
-        return -EOPNOTSUPP;
+       /* Nodes with small feet have little entropy
+        * the NID for this node gives the most entropy in the low bits */
+       while ((ni = lnet_get_next_ni_locked(NULL, ni)))
+               add_device_randomness(&ni->ni_nid, sizeof(ni->ni_nid));
+
+       seeded = 1;
+       return;
 }
 
-void
-lnet_notify_locked (lnet_peer_t *lp, int notifylnd, int alive, time_t when)
+/* NB expects LNET_LOCK held */
+static void
+lnet_add_route_to_rnet(struct lnet_remotenet *rnet, struct lnet_route *route)
 {
-        return;
+       unsigned int len = 0;
+       unsigned int offset = 0;
+       struct list_head *e;
+
+       lnet_shuffle_seed();
+
+       list_for_each(e, &rnet->lrn_routes)
+               len++;
+
+       /*
+        * Randomly adding routes to the list is done to ensure that when
+        * different nodes are using the same list of routers, they end up
+        * preferring different routers.
+        */
+       offset = cfs_rand() % (len + 1);
+       list_for_each(e, &rnet->lrn_routes) {
+               if (offset == 0)
+                       break;
+               offset--;
+       }
+       list_add(&route->lr_list, e);
+       /*
+        * force a router check on the gateway to make sure the route is
+        * alive
+        */
+       route->lr_gateway->lp_rtrcheck_timestamp = 0;
+
+       the_lnet.ln_remote_nets_version++;
+
+       /* add the route on the gateway list */
+       list_add(&route->lr_gwlist, &route->lr_gateway->lp_routes);
+
+       /* take a router reference count on the gateway */
+       lnet_rtr_addref_locked(route->lr_gateway);
 }
 
-#endif
-
-static void
-lnet_rtr_addref_locked(lnet_peer_t *lp)
+int
+lnet_add_route(__u32 net, __u32 hops, lnet_nid_t gateway,
+              unsigned int priority)
 {
-        LASSERT (lp->lp_refcount > 0);
-        LASSERT (lp->lp_rtr_refcount >= 0);
-
-        lp->lp_rtr_refcount++;
-        if (lp->lp_rtr_refcount == 1) {
-                struct list_head *pos;
-
-                /* a simple insertion sort */
-                list_for_each_prev(pos, &the_lnet.ln_routers) {
-                        lnet_peer_t *rtr = list_entry(pos, lnet_peer_t, 
-                                                      lp_rtr_list);
-
-                        if (rtr->lp_nid < lp->lp_nid)
-                                break;
-                }
-
-                list_add(&lp->lp_rtr_list, pos);
-                /* addref for the_lnet.ln_routers */
-                lnet_peer_addref_locked(lp);
-                the_lnet.ln_routers_version++;
-        }
+       struct list_head *route_entry;
+       struct lnet_remotenet *rnet;
+       struct lnet_remotenet *rnet2;
+       struct lnet_route *route;
+       struct lnet_peer_ni *lpni;
+       struct lnet_peer *gw;
+       int add_route;
+       int rc;
+
+       CDEBUG(D_NET, "Add route: remote net %s hops %d priority %u gw %s\n",
+              libcfs_net2str(net), hops, priority, libcfs_nid2str(gateway));
+
+       if (gateway == LNET_NID_ANY ||
+           LNET_NETTYP(LNET_NIDNET(gateway)) == LOLND ||
+           net == LNET_NIDNET(LNET_NID_ANY) ||
+           LNET_NETTYP(net) == LOLND ||
+           LNET_NIDNET(gateway) == net ||
+           (hops != LNET_UNDEFINED_HOPS && (hops < 1 || hops > 255)))
+               return -EINVAL;
+
+       /* it's a local network */
+       if (lnet_islocalnet(net))
+               return -EEXIST;
+
+       /* Assume net, route, all new */
+       LIBCFS_ALLOC(route, sizeof(*route));
+       LIBCFS_ALLOC(rnet, sizeof(*rnet));
+       if (route == NULL || rnet == NULL) {
+               CERROR("Out of memory creating route %s %d %s\n",
+                      libcfs_net2str(net), hops, libcfs_nid2str(gateway));
+               if (route != NULL)
+                       LIBCFS_FREE(route, sizeof(*route));
+               if (rnet != NULL)
+                       LIBCFS_FREE(rnet, sizeof(*rnet));
+               return -ENOMEM;
+       }
+
+       INIT_LIST_HEAD(&rnet->lrn_routes);
+       rnet->lrn_net = net;
+       /* store the local and remote net that the route represents */
+       route->lr_lnet = LNET_NIDNET(gateway);
+       route->lr_net = net;
+       route->lr_priority = priority;
+       route->lr_hops = hops;
+
+       lnet_net_lock(LNET_LOCK_EX);
+
+       /*
+        * lnet_nid2peerni_ex() grabs a ref on the lpni. We will need to
+        * lose that once we're done
+        */
+       lpni = lnet_nid2peerni_ex(gateway, LNET_LOCK_EX);
+       if (IS_ERR(lpni)) {
+               lnet_net_unlock(LNET_LOCK_EX);
+
+               LIBCFS_FREE(route, sizeof(*route));
+               LIBCFS_FREE(rnet, sizeof(*rnet));
+
+               rc = PTR_ERR(lpni);
+               CERROR("Error %d creating route %s %d %s\n", rc,
+                       libcfs_net2str(net), hops,
+                       libcfs_nid2str(gateway));
+               return rc;
+       }
+
+       LASSERT(lpni->lpni_peer_net && lpni->lpni_peer_net->lpn_peer);
+       gw = lpni->lpni_peer_net->lpn_peer;
+
+       route->lr_gateway = gw;
+
+       rnet2 = lnet_find_rnet_locked(net);
+       if (rnet2 == NULL) {
+               /* new network */
+               list_add_tail(&rnet->lrn_list, lnet_net2rnethash(net));
+               rnet2 = rnet;
+       }
+
+       /* Search for a duplicate route (it's a NOOP if it is) */
+       add_route = 1;
+       list_for_each(route_entry, &rnet2->lrn_routes) {
+               struct lnet_route *route2;
+
+               route2 = list_entry(route_entry, struct lnet_route, lr_list);
+               if (route2->lr_gateway == route->lr_gateway) {
+                       add_route = 0;
+                       break;
+               }
+
+               /* our lookups must be true */
+               LASSERT(route2->lr_gateway->lp_primary_nid != gateway);
+       }
+
+       /*
+        * It is possible to add multiple routes through the same peer,
+        * but it'll be using a different NID of that peer. When the
+        * gateway is discovered, discovery will consolidate the different
+        * peers into one peer. In this case the discovery code will have
+        * to move the routes from the peer that's being deleted to the
+        * consolidated peer lp_routes list
+        */
+       if (add_route)
+               lnet_add_route_to_rnet(rnet2, route);
+
+       /*
+        * get rid of the reference on the lpni.
+        */
+       lnet_peer_ni_decref_locked(lpni);
+       lnet_net_unlock(LNET_LOCK_EX);
+
+       rc = 0;
+
+       if (!add_route) {
+               rc = -EEXIST;
+               LIBCFS_FREE(route, sizeof(*route));
+       }
+
+       if (rnet != rnet2)
+               LIBCFS_FREE(rnet, sizeof(*rnet));
+
+       /* kick start the monitor thread to handle the added route */
+       wake_up(&the_lnet.ln_mt_waitq);
+
+       return rc;
 }
 
 static void
-lnet_rtr_decref_locked(lnet_peer_t *lp)
+lnet_del_route_from_rnet(lnet_nid_t gw_nid, struct list_head *route_list,
+                        struct list_head *zombies)
 {
-        LASSERT (lp->lp_refcount > 0);
-        LASSERT (lp->lp_rtr_refcount > 0);
-
-        lp->lp_rtr_refcount--;
-        if (lp->lp_rtr_refcount == 0) {
-                list_del(&lp->lp_rtr_list);
-                /* decref for the_lnet.ln_routers */
-                lnet_peer_decref_locked(lp);
-                the_lnet.ln_routers_version++;
-        }
+       struct lnet_peer *gateway;
+       struct lnet_route *route;
+       struct lnet_route *tmp;
+
+       list_for_each_entry_safe(route, tmp, route_list, lr_list) {
+               gateway = route->lr_gateway;
+               if (gw_nid != LNET_NID_ANY &&
+                   gw_nid != gateway->lp_primary_nid)
+                       continue;
+
+               /*
+                * move to zombie to delete outside the lock
+                * Note that this function is called with the
+                * ln_api_mutex held as well as the exclusive net
+                * lock. Adding to the remote net list happens
+                * under the same conditions. Same goes for the
+                * gateway router list
+                */
+               list_move(&route->lr_list, zombies);
+               the_lnet.ln_remote_nets_version++;
+
+               list_del(&route->lr_gwlist);
+               lnet_rtr_decref_locked(gateway);
+       }
 }
 
-lnet_remotenet_t *
-lnet_find_net_locked (__u32 net)
+int
+lnet_del_route(__u32 net, lnet_nid_t gw_nid)
 {
-        lnet_remotenet_t *rnet;
-        struct list_head *tmp;
-
-        LASSERT (!the_lnet.ln_shutdown);
-
-        list_for_each (tmp, &the_lnet.ln_remote_nets) {
-                rnet = list_entry(tmp, lnet_remotenet_t, lrn_list);
-
-                if (rnet->lrn_net == net)
-                        return rnet;
-        }
-        return NULL;
+       struct list_head rnet_zombies;
+       struct lnet_remotenet *rnet;
+       struct lnet_remotenet *tmp;
+       struct list_head *rn_list;
+       struct lnet_peer_ni *lpni;
+       struct lnet_route *route;
+       struct list_head zombies;
+       struct lnet_peer *lp;
+       int i = 0;
+
+       INIT_LIST_HEAD(&rnet_zombies);
+       INIT_LIST_HEAD(&zombies);
+
+       CDEBUG(D_NET, "Del route: net %s : gw %s\n",
+              libcfs_net2str(net), libcfs_nid2str(gw_nid));
+
+       /* NB Caller may specify either all routes via the given gateway
+        * or a specific route entry actual NIDs) */
+
+       lnet_net_lock(LNET_LOCK_EX);
+
+       lpni = lnet_find_peer_ni_locked(gw_nid);
+       if (lpni) {
+               lp = lpni->lpni_peer_net->lpn_peer;
+               LASSERT(lp);
+               gw_nid = lp->lp_primary_nid;
+               lnet_peer_ni_decref_locked(lpni);
+       }
+
+       if (net != LNET_NIDNET(LNET_NID_ANY)) {
+               rnet = lnet_find_rnet_locked(net);
+               if (!rnet) {
+                       lnet_net_unlock(LNET_LOCK_EX);
+                       return -ENOENT;
+               }
+               lnet_del_route_from_rnet(gw_nid, &rnet->lrn_routes,
+                                        &zombies);
+               if (list_empty(&rnet->lrn_routes))
+                       list_move(&rnet->lrn_list, &rnet_zombies);
+               goto delete_zombies;
+       }
+
+       for (i = 0; i < LNET_REMOTE_NETS_HASH_SIZE; i++) {
+               rn_list = &the_lnet.ln_remote_nets_hash[i];
+
+               list_for_each_entry_safe(rnet, tmp, rn_list, lrn_list) {
+                       lnet_del_route_from_rnet(gw_nid, &rnet->lrn_routes,
+                                                &zombies);
+                       if (list_empty(&rnet->lrn_routes))
+                               list_move(&rnet->lrn_list, &rnet_zombies);
+               }
+       }
+
+delete_zombies:
+       lnet_net_unlock(LNET_LOCK_EX);
+
+       while (!list_empty(&zombies)) {
+               route = list_first_entry(&zombies, struct lnet_route, lr_list);
+               list_del(&route->lr_list);
+               LIBCFS_FREE(route, sizeof(*route));
+       }
+
+       while (!list_empty(&rnet_zombies)) {
+               rnet = list_first_entry(&rnet_zombies, struct lnet_remotenet,
+                                       lrn_list);
+               list_del(&rnet->lrn_list);
+               LIBCFS_FREE(rnet, sizeof(*rnet));
+       }
+
+       return 0;
 }
 
-int
-lnet_add_route (__u32 net, unsigned int hops, lnet_nid_t gateway)
+void
+lnet_destroy_routes (void)
 {
-        struct list_head     zombies;
-        struct list_head    *e;
-        lnet_remotenet_t    *rnet;
-        lnet_remotenet_t    *rnet2;
-        lnet_route_t        *route;
-        lnet_route_t        *route2;
-        lnet_ni_t           *ni;
-        int                  add_route;
-        int                  rc;
-
-        CDEBUG(D_NET, "Add route: net %s hops %u gw %s\n",
-               libcfs_net2str(net), hops, libcfs_nid2str(gateway));
-
-        if (gateway == LNET_NID_ANY ||
-            LNET_NETTYP(LNET_NIDNET(gateway)) == LOLND ||
-            net == LNET_NIDNET(LNET_NID_ANY) ||
-            LNET_NETTYP(net) == LOLND ||
-            LNET_NIDNET(gateway) == net ||
-            hops < 1 || hops > 255)
-                return (-EINVAL);
-
-        if (lnet_islocalnet(net))               /* it's a local network */
-                return 0;                       /* ignore the route entry */
-
-        /* Assume net, route, all new */
-        LIBCFS_ALLOC(route, sizeof(*route));
-        LIBCFS_ALLOC(rnet, sizeof(*rnet));
-        if (route == NULL || rnet == NULL) {
-                CERROR("Out of memory creating route %s %d %s\n",
-                       libcfs_net2str(net), hops, libcfs_nid2str(gateway));
-                if (route != NULL)
-                        LIBCFS_FREE(route, sizeof(*route));
-                if (rnet != NULL)
-                        LIBCFS_FREE(rnet, sizeof(*rnet));
-                return -ENOMEM;
-        }
-
-        CFS_INIT_LIST_HEAD(&rnet->lrn_routes);
-        rnet->lrn_net = net;
-        rnet->lrn_hops = hops;
-
-        LNET_LOCK();
-
-        rc = lnet_nid2peer_locked(&route->lr_gateway, gateway);
-        if (rc != 0) {
-                LNET_UNLOCK();
-
-                LIBCFS_FREE(route, sizeof(*route));
-                LIBCFS_FREE(rnet, sizeof(*rnet));
-
-                if (rc == -EHOSTUNREACH)        /* gateway is not on a local net */
-                        return 0;               /* ignore the route entry */
-
-                CERROR("Error %d creating route %s %d %s\n", rc,
-                       libcfs_net2str(net), hops, libcfs_nid2str(gateway));
-                return rc;
-        }
-
-        LASSERT (!the_lnet.ln_shutdown);
-        CFS_INIT_LIST_HEAD(&zombies);
-
-        rnet2 = lnet_find_net_locked(net);
-        if (rnet2 == NULL) {
-                /* new network */
-                list_add_tail(&rnet->lrn_list, &the_lnet.ln_remote_nets);
-                rnet2 = rnet;
-        }
-
-        if (hops > rnet2->lrn_hops) {
-                /* New route is longer; ignore it */
-                add_route = 0;
-        } else if (hops < rnet2->lrn_hops) {
-                /* new route supercedes all currently known routes to this
-                 * net */
-                list_add(&zombies, &rnet2->lrn_routes);
-                list_del_init(&rnet2->lrn_routes);
-                add_route = 1;
-        } else {
-                add_route = 1;
-                /* New route has the same hopcount as existing routes; search
-                 * for a duplicate route (it's a NOOP if it is) */
-                list_for_each (e, &rnet2->lrn_routes) {
-                        route2 = list_entry(e, lnet_route_t, lr_list);
-
-                        if (route2->lr_gateway == route->lr_gateway) {
-                                add_route = 0;
-                                break;
-                        }
-
-                        /* our loopups must be true */
-                        LASSERT (route2->lr_gateway->lp_nid != gateway);
-                }
-        }
-
-        if (add_route) {
-                ni = route->lr_gateway->lp_ni;
-                lnet_ni_addref_locked(ni);
-
-                LASSERT (rc == 0);
-                list_add_tail(&route->lr_list, &rnet2->lrn_routes);
-                the_lnet.ln_remote_nets_version++;
-
-                lnet_rtr_addref_locked(route->lr_gateway);
-
-                LNET_UNLOCK();
-
-                /* XXX Assume alive */
-                if (ni->ni_lnd->lnd_notify != NULL)
-                        (ni->ni_lnd->lnd_notify)(ni, gateway, 1);
-
-                lnet_ni_decref(ni);
-        } else {
-                lnet_peer_decref_locked(route->lr_gateway);
-                LNET_UNLOCK();
-                LIBCFS_FREE(route, sizeof(*route));
-        }
-
-        if (rnet != rnet2)
-                LIBCFS_FREE(rnet, sizeof(*rnet));
-
-        while (!list_empty(&zombies)) {
-                route = list_entry(zombies.next, lnet_route_t, lr_list);
-                list_del(&route->lr_list);
-
-                LNET_LOCK();
-                lnet_rtr_decref_locked(route->lr_gateway);
-                lnet_peer_decref_locked(route->lr_gateway);
-                LNET_UNLOCK();
-                LIBCFS_FREE(route, sizeof(*route));
-        }
-
-        return rc;
+       lnet_del_route(LNET_NIDNET(LNET_NID_ANY), LNET_NID_ANY);
 }
 
-int
-lnet_check_routes (void)
+int lnet_get_rtr_pool_cfg(int cpt, struct lnet_ioctl_pool_cfg *pool_cfg)
 {
-        lnet_remotenet_t    *rnet;
-        lnet_route_t        *route;
-        lnet_route_t        *route2;
-        struct list_head    *e1;
-        struct list_head    *e2;
-
-        LNET_LOCK();
-
-        list_for_each (e1, &the_lnet.ln_remote_nets) {
-                rnet = list_entry(e1, lnet_remotenet_t, lrn_list);
-
-                route2 = NULL;
-                list_for_each (e2, &rnet->lrn_routes) {
-                        route = list_entry(e2, lnet_route_t, lr_list);
-
-                        if (route2 == NULL)
-                                route2 = route;
-                        else if (route->lr_gateway->lp_ni !=
-                                 route2->lr_gateway->lp_ni) {
-                                LNET_UNLOCK();
-
-                                CERROR("Routes to %s via %s and %s not supported\n",
-                                       libcfs_net2str(rnet->lrn_net),
-                                       libcfs_nid2str(route->lr_gateway->lp_nid),
-                                       libcfs_nid2str(route2->lr_gateway->lp_nid));
-                                return -EINVAL;
-                        }
-                }
-        }
-
-        LNET_UNLOCK();
-        return 0;
+       struct lnet_rtrbufpool *rbp;
+       int i, rc = -ENOENT, j;
+
+       if (the_lnet.ln_rtrpools == NULL)
+               return rc;
+
+
+       cfs_percpt_for_each(rbp, i, the_lnet.ln_rtrpools) {
+               if (i != cpt)
+                       continue;
+
+               lnet_net_lock(i);
+               for (j = 0; j < LNET_NRBPOOLS; j++) {
+                       pool_cfg->pl_pools[j].pl_npages = rbp[j].rbp_npages;
+                       pool_cfg->pl_pools[j].pl_nbuffers = rbp[j].rbp_nbuffers;
+                       pool_cfg->pl_pools[j].pl_credits = rbp[j].rbp_credits;
+                       pool_cfg->pl_pools[j].pl_mincredits = rbp[j].rbp_mincredits;
+               }
+               lnet_net_unlock(i);
+               rc = 0;
+               break;
+       }
+
+       lnet_net_lock(LNET_LOCK_EX);
+       pool_cfg->pl_routing = the_lnet.ln_routing;
+       lnet_net_unlock(LNET_LOCK_EX);
+
+       return rc;
 }
 
 int
-lnet_del_route (__u32 net, lnet_nid_t gw_nid)
+lnet_get_route(int idx, __u32 *net, __u32 *hops,
+              lnet_nid_t *gateway, __u32 *alive, __u32 *priority)
 {
-        lnet_remotenet_t    *rnet;
-        lnet_route_t        *route;
-        struct list_head    *e1;
-        struct list_head    *e2;
-        int                  rc = -ENOENT;
-
-        CDEBUG(D_NET, "Del route: net %s : gw %s\n",
-               libcfs_net2str(net), libcfs_nid2str(gw_nid));
-
-        /* NB Caller may specify either all routes via the given gateway
-         * or a specific route entry actual NIDs) */
-
- again:
-        LNET_LOCK();
-
-        list_for_each (e1, &the_lnet.ln_remote_nets) {
-                rnet = list_entry(e1, lnet_remotenet_t, lrn_list);
-
-                if (!(net == LNET_NIDNET(LNET_NID_ANY) ||
-                      net == rnet->lrn_net))
-                        continue;
-
-                list_for_each (e2, &rnet->lrn_routes) {
-                        route = list_entry(e2, lnet_route_t, lr_list);
-
-                        if (!(gw_nid == LNET_NID_ANY ||
-                              gw_nid == route->lr_gateway->lp_nid))
-                                continue;
-
-                        list_del(&route->lr_list);
-                        the_lnet.ln_remote_nets_version++;
-
-                        if (list_empty(&rnet->lrn_routes))
-                                list_del(&rnet->lrn_list);
-                        else
-                                rnet = NULL;
-
-                        lnet_rtr_decref_locked(route->lr_gateway);
-                        lnet_peer_decref_locked(route->lr_gateway);
-                        LNET_UNLOCK();
-
-                        LIBCFS_FREE(route, sizeof (*route));
-
-                        if (rnet != NULL)
-                                LIBCFS_FREE(rnet, sizeof(*rnet));
-
-                        rc = 0;
-                        goto again;
-                }
-        }
-
-        LNET_UNLOCK();
-        return rc;
+       struct list_head *e1;
+       struct list_head *e2;
+       struct lnet_remotenet *rnet;
+       struct lnet_route        *route;
+       int               cpt;
+       int               i;
+       struct list_head *rn_list;
+
+       cpt = lnet_net_lock_current();
+
+       for (i = 0; i < LNET_REMOTE_NETS_HASH_SIZE; i++) {
+               rn_list = &the_lnet.ln_remote_nets_hash[i];
+               list_for_each(e1, rn_list) {
+                       rnet = list_entry(e1, struct lnet_remotenet, lrn_list);
+
+                       list_for_each(e2, &rnet->lrn_routes) {
+                               route = list_entry(e2, struct lnet_route,
+                                                  lr_list);
+
+                               if (idx-- == 0) {
+                                       *net      = rnet->lrn_net;
+                                       *hops     = route->lr_hops;
+                                       *priority = route->lr_priority;
+                                       *gateway  = route->lr_gateway->lp_primary_nid;
+                                       *alive    = lnet_is_route_alive(route);
+                                       lnet_net_unlock(cpt);
+                                       return 0;
+                               }
+                       }
+               }
+       }
+
+       lnet_net_unlock(cpt);
+       return -ENOENT;
 }
 
 void
-lnet_destroy_routes (void)
+lnet_swap_pinginfo(struct lnet_ping_buffer *pbuf)
 {
-        lnet_del_route(LNET_NIDNET(LNET_NID_ANY), LNET_NID_ANY);
+       struct lnet_ni_status *stat;
+       int nnis;
+       int i;
+
+       __swab32s(&pbuf->pb_info.pi_magic);
+       __swab32s(&pbuf->pb_info.pi_features);
+       __swab32s(&pbuf->pb_info.pi_pid);
+       __swab32s(&pbuf->pb_info.pi_nnis);
+       nnis = pbuf->pb_info.pi_nnis;
+       if (nnis > pbuf->pb_nnis)
+               nnis = pbuf->pb_nnis;
+       for (i = 0; i < nnis; i++) {
+               stat = &pbuf->pb_info.pi_ni[i];
+               __swab64s(&stat->ns_nid);
+               __swab32s(&stat->ns_status);
+       }
+       return;
 }
 
-int
-lnet_get_route (int idx, __u32 *net, __u32 *hops,
-               lnet_nid_t *gateway, __u32 *alive)
+/**
+ * TODO: re-implement
+ */
+static void
+lnet_parse_rc_info(struct lnet_rc_data *rcd)
 {
-        struct list_head    *e1;
-        struct list_head    *e2;
-        lnet_remotenet_t    *rnet;
-        lnet_route_t        *route;
-
-        LNET_LOCK();
-
-        list_for_each (e1, &the_lnet.ln_remote_nets) {
-                rnet = list_entry(e1, lnet_remotenet_t, lrn_list);
-
-                list_for_each (e2, &rnet->lrn_routes) {
-                        route = list_entry(e2, lnet_route_t, lr_list);
-
-                        if (idx-- == 0) {
-                                *net     = rnet->lrn_net;
-                                *hops    = rnet->lrn_hops;
-                                *gateway = route->lr_gateway->lp_nid;
-                                *alive   = route->lr_gateway->lp_alive;
-                                LNET_UNLOCK();
-                                return 0;
-                        }
-                }
-        }
-
-        LNET_UNLOCK();
-        return -ENOENT;
+       rcd = rcd;
 }
 
-#if defined(__KERNEL__) && defined(LNET_ROUTER)
 static void
-lnet_router_checker_event (lnet_event_t *event)
+lnet_router_checker_event(struct lnet_event *event)
 {
-        /* CAVEAT EMPTOR: I'm called with LNET_LOCKed and I'm not allowed to
-         * drop it (that's how come I see _every_ event, even ones that would
-         * overflow my EQ) */
-        lnet_peer_t   *lp;
-        lnet_nid_t     nid;
-
-        if (event->unlinked) {
-                /* The router checker thread has unlinked the rc_md
-                 * and exited. */
-                LASSERT (the_lnet.ln_rc_state == LNET_RC_STATE_UNLINKING);
-                the_lnet.ln_rc_state = LNET_RC_STATE_UNLINKED;
-                mutex_up(&the_lnet.ln_rc_signal);
-                return;
-        }
-
-        LASSERT (event->type == LNET_EVENT_SEND ||
-                 event->type == LNET_EVENT_REPLY);
-
-        nid = (event->type == LNET_EVENT_SEND) ?
-              event->target.nid : event->initiator.nid;
-
-        lp = lnet_find_peer_locked(nid);
-        if (lp == NULL) {
-                /* router may have been removed */
-                CDEBUG(D_NET, "Router %s not found\n", libcfs_nid2str(nid));
-                return;
-        }
-
-        if (event->type == LNET_EVENT_SEND)     /* re-enable another ping */
-                lp->lp_ping_notsent = 0;
-
-        if (lnet_isrouter(lp) &&                /* ignore if no longer a router */
-            (event->status != 0 ||
-             event->type == LNET_EVENT_REPLY)) {
-
-                /* A successful REPLY means the router is up.  If _any_ comms
-                 * to the router fail I assume it's down (this will happen if
-                 * we ping alive routers to try to detect router death before
-                 * apps get burned). */
-
-                lnet_notify_locked(lp, 1, (event->status == 0),
-                                   cfs_time_current_sec());
-
-                /* The router checker will wake up very shortly and do the
-                 * actual notification.  
-                 * XXX If 'lp' stops being a router before then, it will still
-                 * have the notification pending!!! */
-        }
-
-        /* This decref will NOT drop LNET_LOCK (it had to have 1 ref when it
-         * was in the peer table and I've not dropped the lock, so no-one else
-         * can have reduced the refcount) */
-        LASSERT(lp->lp_refcount > 1);
-
-        lnet_peer_decref_locked(lp);
+       struct lnet_rc_data *rcd = event->md.user_ptr;
+       struct lnet_peer_ni *lp;
+
+       LASSERT(rcd != NULL);
+
+       if (event->unlinked) {
+               LNetInvalidateMDHandle(&rcd->rcd_mdh);
+               return;
+       }
+
+       LASSERT(event->type == LNET_EVENT_SEND ||
+               event->type == LNET_EVENT_REPLY);
+
+       lp = rcd->rcd_gateway;
+       LASSERT(lp != NULL);
+
+        /* NB: it's called with holding lnet_res_lock, we have a few
+         * places need to hold both locks at the same time, please take
+         * care of lock ordering */
+       lnet_net_lock(lp->lpni_cpt);
+       if (!lnet_isrouter(lp) || lp->lpni_rcd != rcd) {
+               /* ignore if no longer a router or rcd is replaced */
+               goto out;
+       }
+
+       if (event->type == LNET_EVENT_SEND) {
+               if (event->status == 0)
+                       goto out;
+       }
+
+       /* LNET_EVENT_REPLY */
+       /* A successful REPLY means the router is up.  If _any_ comms
+        * to the router fail I assume it's down (this will happen if
+        * we ping alive routers to try to detect router death before
+        * apps get burned). */
+
+       lnet_notify_locked(lp, 1, !event->status, ktime_get_seconds());
+       /* The router checker will wake up very shortly and do the
+        * actual notification.
+        * XXX If 'lp' stops being a router before then, it will still
+        * have the notification pending!!! */
+
+       if (avoid_asym_router_failure && event->status == 0)
+               lnet_parse_rc_info(rcd);
+
+ out:
+       lnet_net_unlock(lp->lpni_cpt);
 }
 
-static int
-lnet_router_checker(void *arg)
+static void
+lnet_wait_known_routerstate(void)
 {
-        static lnet_ping_info_t   pinginfo;
-
-        int                  rc;
-        lnet_handle_md_t     mdh;
-        lnet_peer_t         *rtr;
-        lnet_md_t            md = {0};
-        struct list_head    *entry;
-        time_t               now;
-        lnet_process_id_t    rtr_id;
-        int                  secs;
-
-        cfs_daemonize("router_checker");
-        cfs_block_allsigs();
+       struct lnet_peer *rtr;
+       struct list_head *entry;
+       int all_known;
 
-        rtr_id.pid = LUSTRE_SRV_LNET_PID;
-
-        LASSERT (the_lnet.ln_rc_state == LNET_RC_STATE_SHUTDOWN);
-
-        /* initialize md content */
-        md.start     = &pinginfo;
-        md.length    = sizeof(pinginfo);
-        md.threshold = LNET_MD_THRESH_INF;
-        md.max_size  = 0;
-        md.options   = LNET_MD_TRUNCATE,
-        md.user_ptr  = NULL;
-        md.eq_handle = the_lnet.ln_rc_eqh;
-
-        rc = LNetMDBind(md, LNET_UNLINK, &mdh);
-
-        if (rc < 0) {
-                CERROR("Can't bind MD: %d\n", rc);
-                the_lnet.ln_rc_state = rc;
-                mutex_up(&the_lnet.ln_rc_signal);
-                return rc;
-        }
-
-        LASSERT (rc == 0);
-
-        the_lnet.ln_rc_state = LNET_RC_STATE_RUNNING;
-        mutex_up(&the_lnet.ln_rc_signal);       /* let my parent go */
-
-        while (the_lnet.ln_rc_state == LNET_RC_STATE_RUNNING) {
-                __u64 version;
-
-                LNET_LOCK();
-rescan:
-                version = the_lnet.ln_routers_version;
+       LASSERT(the_lnet.ln_mt_state == LNET_MT_STATE_RUNNING);
 
-                list_for_each (entry, &the_lnet.ln_routers) {
-                        rtr = list_entry(entry, lnet_peer_t, lp_rtr_list);
+       for (;;) {
+               int cpt = lnet_net_lock_current();
 
-                        lnet_peer_addref_locked(rtr);
+               all_known = 1;
+               list_for_each(entry, &the_lnet.ln_routers) {
+                       rtr = list_entry(entry, struct lnet_peer,
+                                        lp_rtr_list);
 
-                        now = cfs_time_current_sec();
+                       spin_lock(&rtr->lp_lock);
 
-                        if (rtr->lp_ping_deadline != 0 && /* ping timed out? */
-                            now > rtr->lp_ping_deadline)
-                                lnet_notify_locked(rtr, 1, 0, now);
+                       if ((rtr->lp_state & LNET_PEER_DISCOVERED) == 0) {
+                               all_known = 0;
+                               spin_unlock(&rtr->lp_lock);
+                               break;
+                       }
+                       spin_unlock(&rtr->lp_lock);
+               }
 
-                        LNET_UNLOCK();
+               lnet_net_unlock(cpt);
 
-                        /* Run any outstanding notificiations */
-                        lnet_do_notify(rtr);
+               if (all_known)
+                       return;
 
-                        if (rtr->lp_alive) {
-                                secs = live_router_check_interval;
-                        } else {
-                                secs = dead_router_check_interval;
-                        }
-                        if (secs <= 0)
-                                secs = 0;
-
-                        if (secs != 0 &&
-                            !rtr->lp_ping_notsent &&
-                            now > rtr->lp_ping_timestamp + secs) {
-                                CDEBUG(D_NET, "Check: %s\n",
-                                       libcfs_nid2str(rtr->lp_nid));
-
-                                LNET_LOCK();
-                                rtr_id.nid = rtr->lp_nid;
-                                rtr->lp_ping_notsent = 1;
-                                rtr->lp_ping_timestamp = now;
-
-                                if (rtr->lp_ping_deadline == 0)
-                                        rtr->lp_ping_deadline = 
-                                                now + router_ping_timeout;
-
-                                LNET_UNLOCK();
-
-                                LNetGet(LNET_NID_ANY, mdh, rtr_id,
-                                        LNET_RESERVED_PORTAL,
-                                        LNET_PROTO_PING_MATCHBITS, 0);
-                        }
-
-                        LNET_LOCK();
-                        lnet_peer_decref_locked(rtr);
-
-                        if (version != the_lnet.ln_routers_version) {
-                                /* the routers list has changed */
-                                goto rescan;
-                        }
-                }
-
-                LNET_UNLOCK();
-
-                /* Call cfs_pause() here always adds 1 to load average 
-                 * because kernel counts # active tasks as nr_running 
-                 * + nr_uninterruptible. */
-                cfs_schedule_timeout(CFS_TASK_INTERRUPTIBLE,
-                                     cfs_time_seconds(1));
-        }
+               set_current_state(TASK_UNINTERRUPTIBLE);
+               schedule_timeout(cfs_time_seconds(1));
+       }
+}
 
-        LASSERT (the_lnet.ln_rc_state == LNET_RC_STATE_STOPTHREAD);
-        the_lnet.ln_rc_state = LNET_RC_STATE_UNLINKING;
+/* TODO: reimplement */
+void
+lnet_router_ni_update_locked(struct lnet_peer_ni *gw, __u32 net)
+{
+       struct lnet_route *rte;
+       struct lnet_peer *lp;
+
+       if ((gw->lpni_ping_feats & LNET_PING_FEAT_NI_STATUS) != 0)
+               lp = gw->lpni_peer_net->lpn_peer;
+       else
+               return;
+
+       list_for_each_entry(rte, &lp->lp_routes, lr_gwlist) {
+               if (rte->lr_net == net) {
+                       rte->lr_downis = 0;
+                       break;
+               }
+       }
+}
 
-        rc = LNetMDUnlink(mdh);
-        LASSERT (rc == 0);
+static void
+lnet_update_ni_status_locked(void)
+{
+       struct lnet_ni *ni = NULL;
+       time64_t now;
+       time64_t timeout;
+
+       LASSERT(the_lnet.ln_routing);
+
+       timeout = router_ping_timeout +
+                 MAX(live_router_check_interval, dead_router_check_interval);
+
+       now = ktime_get_real_seconds();
+       while ((ni = lnet_get_next_ni_locked(NULL, ni))) {
+               if (ni->ni_net->net_lnd->lnd_type == LOLND)
+                       continue;
+
+               if (now < ni->ni_last_alive + timeout)
+                       continue;
+
+               lnet_ni_lock(ni);
+               /* re-check with lock */
+               if (now < ni->ni_last_alive + timeout) {
+                       lnet_ni_unlock(ni);
+                       continue;
+               }
+
+               LASSERT(ni->ni_status != NULL);
+
+               if (ni->ni_status->ns_status != LNET_NI_STATUS_DOWN) {
+                       CDEBUG(D_NET, "NI(%s:%lld) status changed to down\n",
+                              libcfs_nid2str(ni->ni_nid), timeout);
+                       /* NB: so far, this is the only place to set
+                        * NI status to "down" */
+                       ni->ni_status->ns_status = LNET_NI_STATUS_DOWN;
+               }
+               lnet_ni_unlock(ni);
+       }
+}
 
-        /* The unlink event callback will signal final completion */
-        return 0;
+int lnet_router_pre_mt_start(void)
+{
+       int rc;
+
+       if (check_routers_before_use &&
+           dead_router_check_interval <= 0) {
+               LCONSOLE_ERROR_MSG(0x10a, "'dead_router_check_interval' must be"
+                                  " set if 'check_routers_before_use' is set"
+                                  "\n");
+               return -EINVAL;
+       }
+
+       rc = LNetEQAlloc(0, lnet_router_checker_event, &the_lnet.ln_rc_eqh);
+       if (rc != 0) {
+               CERROR("Can't allocate EQ(0): %d\n", rc);
+               return -ENOMEM;
+       }
+
+       return 0;
 }
 
+void lnet_router_post_mt_start(void)
+{
+       if (check_routers_before_use) {
+               /* Note that a helpful side-effect of pinging all known routers
+                * at startup is that it makes them drop stale connections they
+                * may have to a previous instance of me. */
+               lnet_wait_known_routerstate();
+       }
+}
 
 void
-lnet_wait_known_routerstate(void)
+lnet_router_cleanup(void)
 {
-        lnet_peer_t         *rtr;
-        struct list_head    *entry;
-        int                  all_known;
-
-        for (;;) {
-                LNET_LOCK();
-
-                all_known = 1;
-                list_for_each (entry, &the_lnet.ln_routers) {
-                        rtr = list_entry(entry, lnet_peer_t, lp_rtr_list);
+       int rc;
 
-                        if (rtr->lp_alive_count == 0) {
-                                all_known = 0;
-                                break;
-                        }
-                }
-
-                LNET_UNLOCK();
+       rc = LNetEQFree(the_lnet.ln_rc_eqh);
+       LASSERT(rc == 0);
+       return;
+}
 
-                if (all_known)
-                        return;
+void
+lnet_prune_rc_data(int wait_unlink)
+{
+       wait_unlink = wait_unlink;
+}
 
-                cfs_pause(cfs_time_seconds(1));
-        }
+/*
+ * This function is called from the monitor thread to check if there are
+ * any active routers that need to be checked.
+ */
+inline bool
+lnet_router_checker_active(void)
+{
+       if (the_lnet.ln_mt_state != LNET_MT_STATE_RUNNING)
+               return true;
+
+       /* Router Checker thread needs to run when routing is enabled in
+        * order to call lnet_update_ni_status_locked() */
+       if (the_lnet.ln_routing)
+               return true;
+
+       /* if there are routers that need to be cleaned up then do so */
+       if (!list_empty(&the_lnet.ln_rcd_deathrow) ||
+           !list_empty(&the_lnet.ln_rcd_zombie))
+               return true;
+
+       return !list_empty(&the_lnet.ln_routers) &&
+               (live_router_check_interval > 0 ||
+                dead_router_check_interval > 0);
 }
 
 void
-lnet_router_checker_stop(void)
+lnet_check_routers(void)
 {
-        int       rc;
+       struct lnet_peer *rtr;
+       struct list_head *entry;
+       __u64   version;
+       int     cpt;
 
-        LASSERT (the_lnet.ln_rc_state == LNET_RC_STATE_RUNNING ||
-                 the_lnet.ln_rc_state == LNET_RC_STATE_SHUTDOWN);
+       cpt = lnet_net_lock_current();
+rescan:
+       version = the_lnet.ln_routers_version;
 
-        if (the_lnet.ln_rc_state == LNET_RC_STATE_SHUTDOWN)
-                return;
+       list_for_each(entry, &the_lnet.ln_routers) {
+               rtr = list_entry(entry, struct lnet_peer,
+                                lp_rtr_list);
 
-        the_lnet.ln_rc_state = LNET_RC_STATE_STOPTHREAD;
-        /* block until event callback signals exit */
-        mutex_down(&the_lnet.ln_rc_signal);
+               /* TODO use discovery to determine if router is alive */
 
-        LASSERT (the_lnet.ln_rc_state == LNET_RC_STATE_UNLINKED);
+               /* NB dropped lock */
+               if (version != the_lnet.ln_routers_version) {
+                       /* the routers list has changed */
+                       goto rescan;
+               }
+       }
 
-        rc = LNetEQFree(the_lnet.ln_rc_eqh);
-        LASSERT (rc == 0);
+       if (the_lnet.ln_routing)
+               lnet_update_ni_status_locked();
 
-        the_lnet.ln_rc_state = LNET_RC_STATE_SHUTDOWN;
-}
+       lnet_net_unlock(cpt);
 
-int
-lnet_router_checker_start(void)
-{
-        int  rc;
-
-        LASSERT (the_lnet.ln_rc_state == LNET_RC_STATE_SHUTDOWN);
-
-        if (check_routers_before_use &&
-            dead_router_check_interval <= 0) {
-                LCONSOLE_ERROR_MSG(0x10a, "'dead_router_check_interval' must be"
-                                   " set if 'check_routers_before_use' is set"
-                                   "\n");
-                return -EINVAL;
-        }
-
-        if (live_router_check_interval <= 0 &&
-            dead_router_check_interval <= 0)
-                return 0;
-
-        init_mutex_locked(&the_lnet.ln_rc_signal);
-
-        /* EQ size doesn't matter; the callback is guaranteed to get every
-         * event */
-        rc = LNetEQAlloc(1, lnet_router_checker_event,
-                         &the_lnet.ln_rc_eqh);
-        if (rc != 0) {
-                CERROR("Can't allocate EQ: %d\n", rc);
-                return -ENOMEM;
-        }
-
-        rc = (int)cfs_kernel_thread(lnet_router_checker, NULL, 0);
-        if (rc < 0) {
-                CERROR("Can't start router checker thread: %d\n", rc);
-                goto failed;
-        }
-
-        mutex_down(&the_lnet.ln_rc_signal);     /* wait for checker to startup */
-
-        rc = the_lnet.ln_rc_state;
-        if (rc < 0) {
-                the_lnet.ln_rc_state = LNET_RC_STATE_SHUTDOWN;
-                goto failed;
-        }
-
-        LASSERT (the_lnet.ln_rc_state == LNET_RC_STATE_RUNNING);
-
-        if (check_routers_before_use) {
-                /* Note that a helpful side-effect of pinging all known routers
-                 * at startup is that it makes them drop stale connections they
-                 * may have to a previous instance of me. */
-                lnet_wait_known_routerstate();
-        }
-
-        return 0;
-
- failed:
-        rc = LNetEQFree(the_lnet.ln_rc_eqh);
-        LASSERT (rc == 0);
-        return rc;
+       lnet_prune_rc_data(0); /* don't wait for UNLINK */
 }
 
 void
-lnet_destroy_rtrbuf(lnet_rtrbuf_t *rb, int npages)
+lnet_destroy_rtrbuf(struct lnet_rtrbuf *rb, int npages)
 {
-        int sz = offsetof(lnet_rtrbuf_t, rb_kiov[npages]);
+       int sz = offsetof(struct lnet_rtrbuf, rb_kiov[npages]);
 
-        while (--npages >= 0)
-                cfs_free_page(rb->rb_kiov[npages].kiov_page);
+       while (--npages >= 0)
+               __free_page(rb->rb_kiov[npages].kiov_page);
 
-        LIBCFS_FREE(rb, sz);
+       LIBCFS_FREE(rb, sz);
 }
 
-lnet_rtrbuf_t *
-lnet_new_rtrbuf(lnet_rtrbufpool_t *rbp)
+static struct lnet_rtrbuf *
+lnet_new_rtrbuf(struct lnet_rtrbufpool *rbp, int cpt)
 {
-        int            npages = rbp->rbp_npages;
-        int            sz = offsetof(lnet_rtrbuf_t, rb_kiov[npages]);
-        struct page   *page;
-        lnet_rtrbuf_t *rb;
-        int            i;
-
-        LIBCFS_ALLOC(rb, sz);
-        if (rb == NULL)
-                return NULL;
-
-        rb->rb_pool = rbp;
-
-        for (i = 0; i < npages; i++) {
-                page = cfs_alloc_page(CFS_ALLOC_ZERO | CFS_ALLOC_STD);
-                if (page == NULL) {
-                        while (--i >= 0)
-                                cfs_free_page(rb->rb_kiov[i].kiov_page);
-
-                        LIBCFS_FREE(rb, sz);
-                        return NULL;
-                }
-
-                rb->rb_kiov[i].kiov_len = CFS_PAGE_SIZE;
-                rb->rb_kiov[i].kiov_offset = 0;
-                rb->rb_kiov[i].kiov_page = page;
-        }
-
-        return rb;
+       int            npages = rbp->rbp_npages;
+       int            sz = offsetof(struct lnet_rtrbuf, rb_kiov[npages]);
+       struct page   *page;
+       struct lnet_rtrbuf *rb;
+       int            i;
+
+       LIBCFS_CPT_ALLOC(rb, lnet_cpt_table(), cpt, sz);
+       if (rb == NULL)
+               return NULL;
+
+       rb->rb_pool = rbp;
+
+       for (i = 0; i < npages; i++) {
+               page = cfs_page_cpt_alloc(lnet_cpt_table(), cpt,
+                                         GFP_KERNEL | __GFP_ZERO);
+               if (page == NULL) {
+                       while (--i >= 0)
+                               __free_page(rb->rb_kiov[i].kiov_page);
+
+                       LIBCFS_FREE(rb, sz);
+                       return NULL;
+               }
+
+               rb->rb_kiov[i].kiov_len = PAGE_SIZE;
+               rb->rb_kiov[i].kiov_offset = 0;
+               rb->rb_kiov[i].kiov_page = page;
+       }
+
+       return rb;
 }
 
-void
-lnet_rtrpool_free_bufs(lnet_rtrbufpool_t *rbp)
+static void
+lnet_rtrpool_free_bufs(struct lnet_rtrbufpool *rbp, int cpt)
 {
-        int            npages = rbp->rbp_npages;
-        int            nbuffers = 0;
-        lnet_rtrbuf_t *rb;
-
-        LASSERT (list_empty(&rbp->rbp_msgs));
-        LASSERT (rbp->rbp_credits == rbp->rbp_nbuffers);
-
-        while (!list_empty(&rbp->rbp_bufs)) {
-                LASSERT (rbp->rbp_credits > 0);
-
-                rb = list_entry(rbp->rbp_bufs.next,
-                                lnet_rtrbuf_t, rb_list);
-                list_del(&rb->rb_list);
-                lnet_destroy_rtrbuf(rb, npages);
-                nbuffers++;
-        }
-
-        LASSERT (rbp->rbp_nbuffers == nbuffers);
-        LASSERT (rbp->rbp_credits == nbuffers);
-
-        rbp->rbp_nbuffers = rbp->rbp_credits = 0;
+       int npages = rbp->rbp_npages;
+       struct lnet_rtrbuf *rb;
+       struct list_head tmp;
+
+       if (rbp->rbp_nbuffers == 0) /* not initialized or already freed */
+               return;
+
+       INIT_LIST_HEAD(&tmp);
+
+       lnet_net_lock(cpt);
+       list_splice_init(&rbp->rbp_msgs, &tmp);
+       lnet_drop_routed_msgs_locked(&tmp, cpt);
+       list_splice_init(&rbp->rbp_bufs, &tmp);
+       rbp->rbp_req_nbuffers = 0;
+       rbp->rbp_nbuffers = rbp->rbp_credits = 0;
+       rbp->rbp_mincredits = 0;
+       lnet_net_unlock(cpt);
+
+       /* Free buffers on the free list. */
+       while (!list_empty(&tmp)) {
+               rb = list_entry(tmp.next, struct lnet_rtrbuf, rb_list);
+               list_del(&rb->rb_list);
+               lnet_destroy_rtrbuf(rb, npages);
+       }
 }
 
-int
-lnet_rtrpool_alloc_bufs(lnet_rtrbufpool_t *rbp, int nbufs)
+static int
+lnet_rtrpool_adjust_bufs(struct lnet_rtrbufpool *rbp, int nbufs, int cpt)
 {
-        lnet_rtrbuf_t *rb;
-        int            i;
-
-        if (rbp->rbp_nbuffers != 0) {
-                LASSERT (rbp->rbp_nbuffers == nbufs);
-                return 0;
-        }
-
-        for (i = 0; i < nbufs; i++) {
-                rb = lnet_new_rtrbuf(rbp);
-
-                if (rb == NULL) {
-                        CERROR("Failed to allocate %d router bufs of %d pages\n",
-                               nbufs, rbp->rbp_npages);
-                        return -ENOMEM;
-                }
-
-                rbp->rbp_nbuffers++;
-                rbp->rbp_credits++;
-                rbp->rbp_mincredits++;
-                list_add(&rb->rb_list, &rbp->rbp_bufs);
-
-                /* No allocation "under fire" */
-                /* Otherwise we'd need code to schedule blocked msgs etc */
-                LASSERT (!the_lnet.ln_routing);
-        }
-
-        LASSERT (rbp->rbp_credits == nbufs);
-        return 0;
+       struct list_head rb_list;
+       struct lnet_rtrbuf *rb;
+       int             num_rb;
+       int             num_buffers = 0;
+       int             old_req_nbufs;
+       int             npages = rbp->rbp_npages;
+
+       lnet_net_lock(cpt);
+       /* If we are called for less buffers than already in the pool, we
+        * just lower the req_nbuffers number and excess buffers will be
+        * thrown away as they are returned to the free list.  Credits
+        * then get adjusted as well.
+        * If we already have enough buffers allocated to serve the
+        * increase requested, then we can treat that the same way as we
+        * do the decrease. */
+       num_rb = nbufs - rbp->rbp_nbuffers;
+       if (nbufs <= rbp->rbp_req_nbuffers || num_rb <= 0) {
+               rbp->rbp_req_nbuffers = nbufs;
+               lnet_net_unlock(cpt);
+               return 0;
+       }
+       /* store the older value of rbp_req_nbuffers and then set it to
+        * the new request to prevent lnet_return_rx_credits_locked() from
+        * freeing buffers that we need to keep around */
+       old_req_nbufs = rbp->rbp_req_nbuffers;
+       rbp->rbp_req_nbuffers = nbufs;
+       lnet_net_unlock(cpt);
+
+       INIT_LIST_HEAD(&rb_list);
+
+       /* allocate the buffers on a local list first.  If all buffers are
+        * allocated successfully then join this list to the rbp buffer
+        * list.  If not then free all allocated buffers. */
+       while (num_rb-- > 0) {
+               rb = lnet_new_rtrbuf(rbp, cpt);
+               if (rb == NULL) {
+                       CERROR("Failed to allocate %d route bufs of %d pages\n",
+                              nbufs, npages);
+
+                       lnet_net_lock(cpt);
+                       rbp->rbp_req_nbuffers = old_req_nbufs;
+                       lnet_net_unlock(cpt);
+
+                       goto failed;
+               }
+
+               list_add(&rb->rb_list, &rb_list);
+               num_buffers++;
+       }
+
+       lnet_net_lock(cpt);
+
+       list_splice_tail(&rb_list, &rbp->rbp_bufs);
+       rbp->rbp_nbuffers += num_buffers;
+       rbp->rbp_credits += num_buffers;
+       rbp->rbp_mincredits = rbp->rbp_credits;
+       /* We need to schedule blocked msg using the newly
+        * added buffers. */
+       while (!list_empty(&rbp->rbp_bufs) &&
+              !list_empty(&rbp->rbp_msgs))
+               lnet_schedule_blocked_locked(rbp);
+
+       lnet_net_unlock(cpt);
+
+       return 0;
+
+failed:
+       while (!list_empty(&rb_list)) {
+               rb = list_entry(rb_list.next, struct lnet_rtrbuf, rb_list);
+               list_del(&rb->rb_list);
+               lnet_destroy_rtrbuf(rb, npages);
+       }
+
+       return -ENOMEM;
 }
 
-void
-lnet_rtrpool_init(lnet_rtrbufpool_t *rbp, int npages)
+static void
+lnet_rtrpool_init(struct lnet_rtrbufpool *rbp, int npages)
 {
-        CFS_INIT_LIST_HEAD(&rbp->rbp_msgs);
-        CFS_INIT_LIST_HEAD(&rbp->rbp_bufs);
+       INIT_LIST_HEAD(&rbp->rbp_msgs);
+       INIT_LIST_HEAD(&rbp->rbp_bufs);
 
-        rbp->rbp_npages = npages;
-        rbp->rbp_credits = 0;
-        rbp->rbp_mincredits = 0;
+       rbp->rbp_npages = npages;
+       rbp->rbp_credits = 0;
+       rbp->rbp_mincredits = 0;
 }
 
 void
-lnet_free_rtrpools(void)
+lnet_rtrpools_free(int keep_pools)
 {
-        lnet_rtrpool_free_bufs(&the_lnet.ln_rtrpools[0]);
-        lnet_rtrpool_free_bufs(&the_lnet.ln_rtrpools[1]);
-        lnet_rtrpool_free_bufs(&the_lnet.ln_rtrpools[2]);
+       struct lnet_rtrbufpool *rtrp;
+       int               i;
+
+       if (the_lnet.ln_rtrpools == NULL) /* uninitialized or freed */
+               return;
+
+       cfs_percpt_for_each(rtrp, i, the_lnet.ln_rtrpools) {
+               lnet_rtrpool_free_bufs(&rtrp[LNET_TINY_BUF_IDX], i);
+               lnet_rtrpool_free_bufs(&rtrp[LNET_SMALL_BUF_IDX], i);
+               lnet_rtrpool_free_bufs(&rtrp[LNET_LARGE_BUF_IDX], i);
+       }
+
+       if (!keep_pools) {
+               cfs_percpt_free(the_lnet.ln_rtrpools);
+               the_lnet.ln_rtrpools = NULL;
+       }
 }
 
-void
-lnet_init_rtrpools(void)
+static int
+lnet_nrb_tiny_calculate(void)
 {
-        int small_pages = 1;
-        int large_pages = (LNET_MTU + CFS_PAGE_SIZE - 1) >> CFS_PAGE_SHIFT;
+       int     nrbs = LNET_NRB_TINY;
 
-        lnet_rtrpool_init(&the_lnet.ln_rtrpools[0], 0);
-        lnet_rtrpool_init(&the_lnet.ln_rtrpools[1], small_pages);
-        lnet_rtrpool_init(&the_lnet.ln_rtrpools[2], large_pages);
-}
+       if (tiny_router_buffers < 0) {
+               LCONSOLE_ERROR_MSG(0x10c,
+                                  "tiny_router_buffers=%d invalid when "
+                                  "routing enabled\n", tiny_router_buffers);
+               return -EINVAL;
+       }
 
+       if (tiny_router_buffers > 0)
+               nrbs = tiny_router_buffers;
 
-int
-lnet_alloc_rtrpools(int im_a_router)
+       nrbs /= LNET_CPT_NUMBER;
+       return max(nrbs, LNET_NRB_TINY_MIN);
+}
+
+static int
+lnet_nrb_small_calculate(void)
 {
-        int       rc;
-
-        if (!strcmp(forwarding, "")) {
-                /* not set either way */
-                if (!im_a_router)
-                        return 0;
-        } else if (!strcmp(forwarding, "disabled")) {
-                /* explicitly disabled */
-                return 0;
-        } else if (!strcmp(forwarding, "enabled")) {
-                /* explicitly enabled */
-        } else {
-                LCONSOLE_ERROR_MSG(0x10b, "'forwarding' not set to either "
-                                   "'enabled' or 'disabled'\n");
-                return -EINVAL;
-        }
-
-        if (tiny_router_buffers <= 0) {
-                LCONSOLE_ERROR_MSG(0x10c, "tiny_router_buffers=%d invalid when "
-                                   "routing enabled\n", tiny_router_buffers);
-                rc = -EINVAL;
-                goto failed;
-        }
-
-        rc = lnet_rtrpool_alloc_bufs(&the_lnet.ln_rtrpools[0],
-                                     tiny_router_buffers);
-        if (rc != 0)
-                goto failed;
-
-        if (small_router_buffers <= 0) {
-                LCONSOLE_ERROR_MSG(0x10d, "small_router_buffers=%d invalid when"
-                                   " routing enabled\n", small_router_buffers);
-                rc = -EINVAL;
-                goto failed;
-        }
-
-        rc = lnet_rtrpool_alloc_bufs(&the_lnet.ln_rtrpools[1],
-                                     small_router_buffers);
-        if (rc != 0)
-                goto failed;
-
-        if (large_router_buffers <= 0) {
-                LCONSOLE_ERROR_MSG(0x10e, "large_router_buffers=%d invalid when"
-                                   " routing enabled\n", large_router_buffers);
-                rc = -EINVAL;
-                goto failed;
-        }
-
-        rc = lnet_rtrpool_alloc_bufs(&the_lnet.ln_rtrpools[2],
-                                     large_router_buffers);
-        if (rc != 0)
-                goto failed;
-
-        LNET_LOCK();
-        the_lnet.ln_routing = 1;
-        LNET_UNLOCK();
-
-        return 0;
+       int     nrbs = LNET_NRB_SMALL;
 
- failed:
-        lnet_free_rtrpools();
-        return rc;
-}
+       if (small_router_buffers < 0) {
+               LCONSOLE_ERROR_MSG(0x10c,
+                                  "small_router_buffers=%d invalid when "
+                                  "routing enabled\n", small_router_buffers);
+               return -EINVAL;
+       }
 
-#else
+       if (small_router_buffers > 0)
+               nrbs = small_router_buffers;
 
-int
-lnet_peers_start_down(void)
+       nrbs /= LNET_CPT_NUMBER;
+       return max(nrbs, LNET_NRB_SMALL_MIN);
+}
+
+static int
+lnet_nrb_large_calculate(void)
 {
-        return 0;
+       int     nrbs = LNET_NRB_LARGE;
+
+       if (large_router_buffers < 0) {
+               LCONSOLE_ERROR_MSG(0x10c,
+                                  "large_router_buffers=%d invalid when "
+                                  "routing enabled\n", large_router_buffers);
+               return -EINVAL;
+       }
+
+       if (large_router_buffers > 0)
+               nrbs = large_router_buffers;
+
+       nrbs /= LNET_CPT_NUMBER;
+       return max(nrbs, LNET_NRB_LARGE_MIN);
 }
 
 int
-lnet_peer_buffer_credits(lnet_ni_t *ni)
+lnet_rtrpools_alloc(int im_a_router)
 {
-        return 0;
+       struct lnet_rtrbufpool *rtrp;
+       int     nrb_tiny;
+       int     nrb_small;
+       int     nrb_large;
+       int     rc;
+       int     i;
+
+       if (!strcmp(forwarding, "")) {
+               /* not set either way */
+               if (!im_a_router)
+                       return 0;
+       } else if (!strcmp(forwarding, "disabled")) {
+               /* explicitly disabled */
+               return 0;
+       } else if (!strcmp(forwarding, "enabled")) {
+               /* explicitly enabled */
+       } else {
+               LCONSOLE_ERROR_MSG(0x10b, "'forwarding' not set to either "
+                                  "'enabled' or 'disabled'\n");
+               return -EINVAL;
+       }
+
+       nrb_tiny = lnet_nrb_tiny_calculate();
+       if (nrb_tiny < 0)
+               return -EINVAL;
+
+       nrb_small = lnet_nrb_small_calculate();
+       if (nrb_small < 0)
+               return -EINVAL;
+
+       nrb_large = lnet_nrb_large_calculate();
+       if (nrb_large < 0)
+               return -EINVAL;
+
+       the_lnet.ln_rtrpools = cfs_percpt_alloc(lnet_cpt_table(),
+                                               LNET_NRBPOOLS *
+                                               sizeof(struct lnet_rtrbufpool));
+       if (the_lnet.ln_rtrpools == NULL) {
+               LCONSOLE_ERROR_MSG(0x10c,
+                                  "Failed to initialize router buffe pool\n");
+               return -ENOMEM;
+       }
+
+       cfs_percpt_for_each(rtrp, i, the_lnet.ln_rtrpools) {
+               lnet_rtrpool_init(&rtrp[LNET_TINY_BUF_IDX], 0);
+               rc = lnet_rtrpool_adjust_bufs(&rtrp[LNET_TINY_BUF_IDX],
+                                             nrb_tiny, i);
+               if (rc != 0)
+                       goto failed;
+
+               lnet_rtrpool_init(&rtrp[LNET_SMALL_BUF_IDX],
+                                 LNET_NRB_SMALL_PAGES);
+               rc = lnet_rtrpool_adjust_bufs(&rtrp[LNET_SMALL_BUF_IDX],
+                                             nrb_small, i);
+               if (rc != 0)
+                       goto failed;
+
+               lnet_rtrpool_init(&rtrp[LNET_LARGE_BUF_IDX],
+                                 LNET_NRB_LARGE_PAGES);
+               rc = lnet_rtrpool_adjust_bufs(&rtrp[LNET_LARGE_BUF_IDX],
+                                             nrb_large, i);
+               if (rc != 0)
+                       goto failed;
+       }
+
+       lnet_net_lock(LNET_LOCK_EX);
+       the_lnet.ln_routing = 1;
+       lnet_net_unlock(LNET_LOCK_EX);
+       wake_up(&the_lnet.ln_mt_waitq);
+       return 0;
+
+ failed:
+       lnet_rtrpools_free(0);
+       return rc;
 }
 
-void
-lnet_router_checker_stop(void)
+static int
+lnet_rtrpools_adjust_helper(int tiny, int small, int large)
 {
-        return;
+       int nrb = 0;
+       int rc = 0;
+       int i;
+       struct lnet_rtrbufpool *rtrp;
+
+       /* If the provided values for each buffer pool are different than the
+        * configured values, we need to take action. */
+       if (tiny >= 0) {
+               tiny_router_buffers = tiny;
+               nrb = lnet_nrb_tiny_calculate();
+               cfs_percpt_for_each(rtrp, i, the_lnet.ln_rtrpools) {
+                       rc = lnet_rtrpool_adjust_bufs(&rtrp[LNET_TINY_BUF_IDX],
+                                                     nrb, i);
+                       if (rc != 0)
+                               return rc;
+               }
+       }
+       if (small >= 0) {
+               small_router_buffers = small;
+               nrb = lnet_nrb_small_calculate();
+               cfs_percpt_for_each(rtrp, i, the_lnet.ln_rtrpools) {
+                       rc = lnet_rtrpool_adjust_bufs(&rtrp[LNET_SMALL_BUF_IDX],
+                                                     nrb, i);
+                       if (rc != 0)
+                               return rc;
+               }
+       }
+       if (large >= 0) {
+               large_router_buffers = large;
+               nrb = lnet_nrb_large_calculate();
+               cfs_percpt_for_each(rtrp, i, the_lnet.ln_rtrpools) {
+                       rc = lnet_rtrpool_adjust_bufs(&rtrp[LNET_LARGE_BUF_IDX],
+                                                     nrb, i);
+                       if (rc != 0)
+                               return rc;
+               }
+       }
+
+       return 0;
 }
 
 int
-lnet_router_checker_start(void)
+lnet_rtrpools_adjust(int tiny, int small, int large)
 {
-        return 0;
+       /* this function doesn't revert the changes if adding new buffers
+        * failed.  It's up to the user space caller to revert the
+        * changes. */
+
+       if (!the_lnet.ln_routing)
+               return 0;
+
+       return lnet_rtrpools_adjust_helper(tiny, small, large);
 }
 
-void
-lnet_free_rtrpools (void)
+int
+lnet_rtrpools_enable(void)
 {
+       int rc = 0;
+
+       if (the_lnet.ln_routing)
+               return 0;
+
+       if (the_lnet.ln_rtrpools == NULL)
+               /* If routing is turned off, and we have never
+                * initialized the pools before, just call the
+                * standard buffer pool allocation routine as
+                * if we are just configuring this for the first
+                * time. */
+               rc = lnet_rtrpools_alloc(1);
+       else
+               rc = lnet_rtrpools_adjust_helper(0, 0, 0);
+       if (rc != 0)
+               return rc;
+
+       lnet_net_lock(LNET_LOCK_EX);
+       the_lnet.ln_routing = 1;
+
+       the_lnet.ln_ping_target->pb_info.pi_features &=
+               ~LNET_PING_FEAT_RTE_DISABLED;
+       lnet_net_unlock(LNET_LOCK_EX);
+
+       return rc;
 }
 
 void
-lnet_init_rtrpools (void)
+lnet_rtrpools_disable(void)
 {
+       if (!the_lnet.ln_routing)
+               return;
+
+       lnet_net_lock(LNET_LOCK_EX);
+       the_lnet.ln_routing = 0;
+       the_lnet.ln_ping_target->pb_info.pi_features |=
+               LNET_PING_FEAT_RTE_DISABLED;
+
+       tiny_router_buffers = 0;
+       small_router_buffers = 0;
+       large_router_buffers = 0;
+       lnet_net_unlock(LNET_LOCK_EX);
+       lnet_rtrpools_free(1);
 }
 
 int
-lnet_alloc_rtrpools (int im_a_arouter)
+lnet_notify(struct lnet_ni *ni, lnet_nid_t nid, int alive, time64_t when)
 {
-        return 0;
+       struct lnet_peer_ni *lp = NULL;
+       time64_t now = ktime_get_seconds();
+       int cpt = lnet_cpt_of_nid(nid, ni);
+
+       LASSERT (!in_interrupt ());
+
+       CDEBUG (D_NET, "%s notifying %s: %s\n",
+               (ni == NULL) ? "userspace" : libcfs_nid2str(ni->ni_nid),
+               libcfs_nid2str(nid),
+               alive ? "up" : "down");
+
+       if (ni != NULL &&
+           LNET_NIDNET(ni->ni_nid) != LNET_NIDNET(nid)) {
+               CWARN("Ignoring notification of %s %s by %s (different net)\n",
+                     libcfs_nid2str(nid), alive ? "birth" : "death",
+                     libcfs_nid2str(ni->ni_nid));
+               return -EINVAL;
+       }
+
+       /* can't do predictions... */
+       if (when > now) {
+               CWARN("Ignoring prediction from %s of %s %s "
+                     "%lld seconds in the future\n",
+                     (ni == NULL) ? "userspace" : libcfs_nid2str(ni->ni_nid),
+                     libcfs_nid2str(nid), alive ? "up" : "down", when - now);
+               return -EINVAL;
+       }
+
+       if (ni != NULL && !alive &&             /* LND telling me she's down */
+           !auto_down) {                       /* auto-down disabled */
+               CDEBUG(D_NET, "Auto-down disabled\n");
+               return 0;
+       }
+
+       lnet_net_lock(cpt);
+
+       if (the_lnet.ln_state != LNET_STATE_RUNNING) {
+               lnet_net_unlock(cpt);
+               return -ESHUTDOWN;
+       }
+
+       lp = lnet_find_peer_ni_locked(nid);
+       if (lp == NULL) {
+               /* nid not found */
+               lnet_net_unlock(cpt);
+               CDEBUG(D_NET, "%s not found\n", libcfs_nid2str(nid));
+               return 0;
+       }
+
+       /*
+        * It is possible for this function to be called for the same peer
+        * but with different NIs. We want to synchronize the notification
+        * between the different calls. So we will use the lpni_cpt to
+        * grab the net lock.
+        */
+       if (lp->lpni_cpt != cpt) {
+               lnet_net_unlock(cpt);
+               cpt = lp->lpni_cpt;
+               lnet_net_lock(cpt);
+       }
+
+       /* We can't fully trust LND on reporting exact peer last_alive
+        * if he notifies us about dead peer. For example ksocklnd can
+        * call us with when == _time_when_the_node_was_booted_ if
+        * no connections were successfully established */
+       if (ni != NULL && !alive && when < lp->lpni_last_alive)
+               when = lp->lpni_last_alive;
+
+       lnet_notify_locked(lp, ni == NULL, alive, when);
+
+       if (ni != NULL)
+               lnet_ni_notify_locked(ni, lp);
+
+       lnet_peer_ni_decref_locked(lp);
+
+       lnet_net_unlock(cpt);
+       return 0;
 }
-
-#endif
+EXPORT_SYMBOL(lnet_notify);