/* -*- mode: c; c-basic-offset: 8; indent-tabs-mode: nil; -*-
* vim:expandtab:shiftwidth=8:tabstop=8:
*
- * api/api-ni.c
- * Network Interface code
- *
* Copyright (c) 2001-2003 Cluster File Systems, Inc.
- * Copyright (c) 2001-2002 Sandia National Laboratories
*
* This file is part of Lustre, http://www.sf.net/projects/lustre/
*
* Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
*/
-#include <portals/api-support.h>
+#define DEBUG_SUBSYSTEM S_LNET
+#include <lnet/lib-lnet.h>
+
+#ifdef __KERNEL__
+#define D_LNI D_CONSOLE
+#else
+#define D_LNI D_CONFIG
+#endif
+
+lnet_t the_lnet; /* THE state of the network */
+
+#ifdef __KERNEL__
-/* Put some magic in the NI handle so uninitialised/zeroed handles are easy
- * to spot */
-#define NI_HANDLE_MAGIC 0xebc0de00
-#define NI_HANDLE_MASK 0x000000ff
-#define MAX_NIS 8
-static nal_t *ptl_interfaces[MAX_NIS];
-int ptl_num_interfaces = 0;
+static char *ip2nets = "";
+CFS_MODULE_PARM(ip2nets, "s", charp, 0444,
+ "LNET network <- IP table");
-nal_t *ptl_hndl2nal(ptl_handle_any_t *handle)
+static char *networks = "";
+CFS_MODULE_PARM(networks, "s", charp, 0444,
+ "local networks");
+
+static char *routes = "";
+CFS_MODULE_PARM(routes, "s", charp, 0444,
+ "routes to non-local networks");
+
+static char *portals_compatibility = "none";
+CFS_MODULE_PARM(portals_compatibility, "s", charp, 0444,
+ "wire protocol compatibility: 'strong'|'weak'|'none'");
+
+char *
+lnet_get_routes(void)
{
- unsigned int idx = handle->nal_idx;
+ return routes;
+}
- /* XXX we really rely on the caller NOT racing with interface
- * setup/teardown. That ensures her NI handle can't get
- * invalidated out from under her (or worse, swapped for a
- * completely different interface!) */
+char *
+lnet_get_networks(void)
+{
+ char *nets;
+ int rc;
- if (((idx ^ NI_HANDLE_MAGIC) & ~NI_HANDLE_MASK) != 0)
+ if (*networks != 0 && *ip2nets != 0) {
+ LCONSOLE_ERROR_MSG(0x101, "Please specify EITHER 'networks' or "
+ "'ip2nets' but not both at once\n");
return NULL;
+ }
+
+ if (*ip2nets != 0) {
+ rc = lnet_parse_ip2nets(&nets, ip2nets);
+ return (rc == 0) ? nets : NULL;
+ }
- idx &= NI_HANDLE_MASK;
- if (idx < MAX_NIS)
- return ptl_interfaces[idx];
+ if (*networks != 0)
+ return networks;
- return NULL;
+ return "tcp";
+}
+
+int
+lnet_get_portals_compatibility(void)
+{
+ if (!strcmp(portals_compatibility, "none")) {
+ return 0;
+ }
+
+ if (!strcmp(portals_compatibility, "weak")) {
+ return 1;
+ LCONSOLE_WARN("Starting in weak portals-compatible mode\n");
+ }
+
+ if (!strcmp(portals_compatibility, "strong")) {
+ return 2;
+ LCONSOLE_WARN("Starting in strong portals-compatible mode\n");
+ }
+
+ LCONSOLE_ERROR_MSG(0x102, "portals_compatibility=\"%s\" not supported\n",
+ portals_compatibility);
+ return -EINVAL;
}
-int ptl_ni_init(void)
+void
+lnet_init_locks(void)
{
- int i;
+ spin_lock_init (&the_lnet.ln_lock);
+ cfs_waitq_init (&the_lnet.ln_waitq);
+ init_mutex(&the_lnet.ln_lnd_mutex);
+ init_mutex(&the_lnet.ln_api_mutex);
+}
+
+void
+lnet_fini_locks(void)
+{
+}
- LASSERT (MAX_NIS <= (NI_HANDLE_MASK + 1));
+#else
+
+char *
+lnet_get_routes(void)
+{
+ char *str = getenv("LNET_ROUTES");
- for (i = 0; i < MAX_NIS; i++)
- ptl_interfaces[i] = NULL;
+ return (str == NULL) ? "" : str;
+}
- return PTL_OK;
+char *
+lnet_get_networks (void)
+{
+ static char default_networks[256];
+ char *networks = getenv ("LNET_NETWORKS");
+ char *ip2nets = getenv ("LNET_IP2NETS");
+ char *str;
+ char *sep;
+ int len;
+ int nob;
+ int rc;
+ struct list_head *tmp;
+
+#ifdef NOT_YET
+ if (networks != NULL && ip2nets != NULL) {
+ LCONSOLE_ERROR_MSG(0x103, "Please set EITHER 'LNET_NETWORKS' or"
+ " 'LNET_IP2NETS' but not both at once\n");
+ return NULL;
+ }
+
+ if (ip2nets != NULL) {
+ rc = lnet_parse_ip2nets(&networks, ip2nets);
+ return (rc == 0) ? networks : NULL;
+ }
+#else
+ ip2nets = NULL;
+ rc = 0;
+#endif
+ if (networks != NULL)
+ return networks;
+
+ /* In userland, the default 'networks=' is the list of known net types */
+
+ len = sizeof(default_networks);
+ str = default_networks;
+ *str = 0;
+ sep = "";
+
+ list_for_each (tmp, &the_lnet.ln_lnds) {
+ lnd_t *lnd = list_entry(tmp, lnd_t, lnd_list);
+
+ nob = snprintf(str, len, "%s%s", sep,
+ libcfs_lnd2str(lnd->lnd_type));
+ len -= nob;
+ if (len < 0) {
+ /* overflowed the string; leave it where it was */
+ *str = 0;
+ break;
+ }
+
+ str += nob;
+ sep = ",";
+ }
+
+ return default_networks;
}
-void ptl_ni_fini(void)
+int
+lnet_get_portals_compatibility(void)
{
- int i;
+ return 0;
+}
- for (i = 0; i < MAX_NIS; i++) {
- nal_t *nal = ptl_interfaces[i];
- if (!nal)
- continue;
+# ifndef HAVE_LIBPTHREAD
+
+void lnet_init_locks(void)
+{
+ the_lnet.ln_lock = 0;
+ the_lnet.ln_lnd_mutex = 0;
+ the_lnet.ln_api_mutex = 0;
+}
+
+void lnet_fini_locks(void)
+{
+ LASSERT (the_lnet.ln_api_mutex == 0);
+ LASSERT (the_lnet.ln_lnd_mutex == 0);
+ LASSERT (the_lnet.ln_lock == 0);
+}
+
+# else
+
+void lnet_init_locks(void)
+{
+ pthread_cond_init(&the_lnet.ln_cond, NULL);
+ pthread_mutex_init(&the_lnet.ln_lock, NULL);
+ pthread_mutex_init(&the_lnet.ln_lnd_mutex, NULL);
+ pthread_mutex_init(&the_lnet.ln_api_mutex, NULL);
+}
+
+void lnet_fini_locks(void)
+{
+ pthread_mutex_destroy(&the_lnet.ln_api_mutex);
+ pthread_mutex_destroy(&the_lnet.ln_lnd_mutex);
+ pthread_mutex_destroy(&the_lnet.ln_lock);
+ pthread_cond_destroy(&the_lnet.ln_cond);
+}
+
+# endif
+#endif
+
+void lnet_assert_wire_constants (void)
+{
+ /* Wire protocol assertions generated by 'wirecheck'
+ * running on Linux robert.bartonsoftware.com 2.6.8-1.521
+ * #1 Mon Aug 16 09:01:18 EDT 2004 i686 athlon i386 GNU/Linux
+ * with gcc version 3.3.3 20040412 (Red Hat Linux 3.3.3-7) */
+
+ /* Constants... */
+ CLASSERT (LNET_PROTO_TCP_MAGIC == 0xeebc0ded);
+ CLASSERT (LNET_PROTO_TCP_VERSION_MAJOR == 1);
+ CLASSERT (LNET_PROTO_TCP_VERSION_MINOR == 0);
+ CLASSERT (LNET_MSG_ACK == 0);
+ CLASSERT (LNET_MSG_PUT == 1);
+ CLASSERT (LNET_MSG_GET == 2);
+ CLASSERT (LNET_MSG_REPLY == 3);
+ CLASSERT (LNET_MSG_HELLO == 4);
+
+ /* Checks for struct ptl_handle_wire_t */
+ CLASSERT ((int)sizeof(lnet_handle_wire_t) == 16);
+ CLASSERT ((int)offsetof(lnet_handle_wire_t, wh_interface_cookie) == 0);
+ CLASSERT ((int)sizeof(((lnet_handle_wire_t *)0)->wh_interface_cookie) == 8);
+ CLASSERT ((int)offsetof(lnet_handle_wire_t, wh_object_cookie) == 8);
+ CLASSERT ((int)sizeof(((lnet_handle_wire_t *)0)->wh_object_cookie) == 8);
- if (nal->shutdown)
- nal->shutdown(nal, i);
+ /* Checks for struct lnet_magicversion_t */
+ CLASSERT ((int)sizeof(lnet_magicversion_t) == 8);
+ CLASSERT ((int)offsetof(lnet_magicversion_t, magic) == 0);
+ CLASSERT ((int)sizeof(((lnet_magicversion_t *)0)->magic) == 4);
+ CLASSERT ((int)offsetof(lnet_magicversion_t, version_major) == 4);
+ CLASSERT ((int)sizeof(((lnet_magicversion_t *)0)->version_major) == 2);
+ CLASSERT ((int)offsetof(lnet_magicversion_t, version_minor) == 6);
+ CLASSERT ((int)sizeof(((lnet_magicversion_t *)0)->version_minor) == 2);
+
+ /* Checks for struct lnet_hdr_t */
+ CLASSERT ((int)sizeof(lnet_hdr_t) == 72);
+ CLASSERT ((int)offsetof(lnet_hdr_t, dest_nid) == 0);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->dest_nid) == 8);
+ CLASSERT ((int)offsetof(lnet_hdr_t, src_nid) == 8);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->src_nid) == 8);
+ CLASSERT ((int)offsetof(lnet_hdr_t, dest_pid) == 16);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->dest_pid) == 4);
+ CLASSERT ((int)offsetof(lnet_hdr_t, src_pid) == 20);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->src_pid) == 4);
+ CLASSERT ((int)offsetof(lnet_hdr_t, type) == 24);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->type) == 4);
+ CLASSERT ((int)offsetof(lnet_hdr_t, payload_length) == 28);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->payload_length) == 4);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg) == 32);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg) == 40);
+
+ /* Ack */
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.ack.dst_wmd) == 32);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.ack.dst_wmd) == 16);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.ack.match_bits) == 48);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.ack.match_bits) == 8);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.ack.mlength) == 56);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.ack.mlength) == 4);
+
+ /* Put */
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.put.ack_wmd) == 32);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.put.ack_wmd) == 16);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.put.match_bits) == 48);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.put.match_bits) == 8);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.put.hdr_data) == 56);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.put.hdr_data) == 8);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.put.ptl_index) == 64);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.put.ptl_index) == 4);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.put.offset) == 68);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.put.offset) == 4);
+
+ /* Get */
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.get.return_wmd) == 32);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.get.return_wmd) == 16);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.get.match_bits) == 48);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.get.match_bits) == 8);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.get.ptl_index) == 56);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.get.ptl_index) == 4);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.get.src_offset) == 60);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.get.src_offset) == 4);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.get.sink_length) == 64);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.get.sink_length) == 4);
+
+ /* Reply */
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.reply.dst_wmd) == 32);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.reply.dst_wmd) == 16);
+
+ /* Hello */
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.hello.incarnation) == 32);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.hello.incarnation) == 8);
+ CLASSERT ((int)offsetof(lnet_hdr_t, msg.hello.type) == 40);
+ CLASSERT ((int)sizeof(((lnet_hdr_t *)0)->msg.hello.type) == 4);
+}
+
+lnd_t *
+lnet_find_lnd_by_type (int type)
+{
+ lnd_t *lnd;
+ struct list_head *tmp;
+
+ /* holding lnd mutex */
+ list_for_each (tmp, &the_lnet.ln_lnds) {
+ lnd = list_entry(tmp, lnd_t, lnd_list);
+
+ if (lnd->lnd_type == type)
+ return lnd;
}
+
+ return NULL;
}
-#ifdef __KERNEL__
-DECLARE_MUTEX(ptl_ni_init_mutex);
+void
+lnet_register_lnd (lnd_t *lnd)
+{
+ LNET_MUTEX_DOWN(&the_lnet.ln_lnd_mutex);
+
+ LASSERT (the_lnet.ln_init);
+ LASSERT (libcfs_isknown_lnd(lnd->lnd_type));
+ LASSERT (lnet_find_lnd_by_type(lnd->lnd_type) == NULL);
+
+ list_add_tail (&lnd->lnd_list, &the_lnet.ln_lnds);
+ lnd->lnd_refcount = 0;
+
+ CDEBUG(D_NET, "%s LND registered\n", libcfs_lnd2str(lnd->lnd_type));
+
+ LNET_MUTEX_UP(&the_lnet.ln_lnd_mutex);
+}
+
+void
+lnet_unregister_lnd (lnd_t *lnd)
+{
+ LNET_MUTEX_DOWN(&the_lnet.ln_lnd_mutex);
+
+ LASSERT (the_lnet.ln_init);
+ LASSERT (lnet_find_lnd_by_type(lnd->lnd_type) == lnd);
+ LASSERT (lnd->lnd_refcount == 0);
+
+ list_del (&lnd->lnd_list);
+ CDEBUG(D_NET, "%s LND unregistered\n", libcfs_lnd2str(lnd->lnd_type));
+
+ LNET_MUTEX_UP(&the_lnet.ln_lnd_mutex);
+}
+
+#ifndef LNET_USE_LIB_FREELIST
+
+int
+lnet_descriptor_setup (void)
+{
+ return 0;
+}
+
+void
+lnet_descriptor_cleanup (void)
+{
+}
+
+#else
+
+int
+lnet_freelist_init (lnet_freelist_t *fl, int n, int size)
+{
+ char *space;
+
+ LASSERT (n > 0);
+
+ size += offsetof (lnet_freeobj_t, fo_contents);
+
+ LIBCFS_ALLOC(space, n * size);
+ if (space == NULL)
+ return (-ENOMEM);
+
+ CFS_INIT_LIST_HEAD (&fl->fl_list);
+ fl->fl_objs = space;
+ fl->fl_nobjs = n;
+ fl->fl_objsize = size;
+
+ do
+ {
+ memset (space, 0, size);
+ list_add ((struct list_head *)space, &fl->fl_list);
+ space += size;
+ } while (--n != 0);
+
+ return (0);
+}
+
+void
+lnet_freelist_fini (lnet_freelist_t *fl)
+{
+ struct list_head *el;
+ int count;
+
+ if (fl->fl_nobjs == 0)
+ return;
+
+ count = 0;
+ for (el = fl->fl_list.next; el != &fl->fl_list; el = el->next)
+ count++;
+
+ LASSERT (count == fl->fl_nobjs);
+
+ LIBCFS_FREE(fl->fl_objs, fl->fl_nobjs * fl->fl_objsize);
+ memset (fl, 0, sizeof (fl));
+}
-static void ptl_ni_init_mutex_enter (void)
+int
+lnet_descriptor_setup (void)
{
- down (&ptl_ni_init_mutex);
+ /* NB on failure caller must still call lnet_descriptor_cleanup */
+ /* ****** */
+ int rc;
+
+ memset (&the_lnet.ln_free_mes, 0, sizeof (the_lnet.ln_free_mes));
+ memset (&the_lnet.ln_free_msgs, 0, sizeof (the_lnet.ln_free_msgs));
+ memset (&the_lnet.ln_free_mds, 0, sizeof (the_lnet.ln_free_mds));
+ memset (&the_lnet.ln_free_eqs, 0, sizeof (the_lnet.ln_free_eqs));
+
+ rc = lnet_freelist_init(&the_lnet.ln_free_mes,
+ MAX_MES, sizeof (lnet_me_t));
+ if (rc != 0)
+ return (rc);
+
+ rc = lnet_freelist_init(&the_lnet.ln_free_msgs,
+ MAX_MSGS, sizeof (lnet_msg_t));
+ if (rc != 0)
+ return (rc);
+
+ rc = lnet_freelist_init(&the_lnet.ln_free_mds,
+ MAX_MDS, sizeof (lnet_libmd_t));
+ if (rc != 0)
+ return (rc);
+
+ rc = lnet_freelist_init(&the_lnet.ln_free_eqs,
+ MAX_EQS, sizeof (lnet_eq_t));
+ return (rc);
}
-static void ptl_ni_init_mutex_exit (void)
+void
+lnet_descriptor_cleanup (void)
{
- up (&ptl_ni_init_mutex);
+ lnet_freelist_fini (&the_lnet.ln_free_mes);
+ lnet_freelist_fini (&the_lnet.ln_free_msgs);
+ lnet_freelist_fini (&the_lnet.ln_free_mds);
+ lnet_freelist_fini (&the_lnet.ln_free_eqs);
}
+#endif
+
+__u64
+lnet_create_interface_cookie (void)
+{
+ /* NB the interface cookie in wire handles guards against delayed
+ * replies and ACKs appearing valid after reboot. Initialisation time,
+ * even if it's only implemented to millisecond resolution is probably
+ * easily good enough. */
+ struct timeval tv;
+ __u64 cookie;
+#ifndef __KERNEL__
+ int rc = gettimeofday (&tv, NULL);
+ LASSERT (rc == 0);
#else
-static void ptl_ni_init_mutex_enter (void)
+ do_gettimeofday(&tv);
+#endif
+ cookie = tv.tv_sec;
+ cookie *= 1000000;
+ cookie += tv.tv_usec;
+ return cookie;
+}
+
+int
+lnet_setup_handle_hash (void)
+{
+ int i;
+
+ /* Arbitrary choice of hash table size */
+#ifdef __KERNEL__
+ the_lnet.ln_lh_hash_size = CFS_PAGE_SIZE / sizeof (struct list_head);
+#else
+ the_lnet.ln_lh_hash_size = (MAX_MES + MAX_MDS + MAX_EQS)/4;
+#endif
+ LIBCFS_ALLOC(the_lnet.ln_lh_hash_table,
+ the_lnet.ln_lh_hash_size * sizeof (struct list_head));
+ if (the_lnet.ln_lh_hash_table == NULL)
+ return (-ENOMEM);
+
+ for (i = 0; i < the_lnet.ln_lh_hash_size; i++)
+ CFS_INIT_LIST_HEAD (&the_lnet.ln_lh_hash_table[i]);
+
+ the_lnet.ln_next_object_cookie = LNET_COOKIE_TYPES;
+
+ return (0);
+}
+
+void
+lnet_cleanup_handle_hash (void)
+{
+ if (the_lnet.ln_lh_hash_table == NULL)
+ return;
+
+ LIBCFS_FREE(the_lnet.ln_lh_hash_table,
+ the_lnet.ln_lh_hash_size * sizeof (struct list_head));
+}
+
+lnet_libhandle_t *
+lnet_lookup_cookie (__u64 cookie, int type)
+{
+ /* ALWAYS called with LNET_LOCK held */
+ struct list_head *list;
+ struct list_head *el;
+ unsigned int hash;
+
+ if ((cookie & (LNET_COOKIE_TYPES - 1)) != type)
+ return (NULL);
+
+ hash = ((unsigned int)cookie) % the_lnet.ln_lh_hash_size;
+ list = &the_lnet.ln_lh_hash_table[hash];
+
+ list_for_each (el, list) {
+ lnet_libhandle_t *lh = list_entry (el, lnet_libhandle_t,
+ lh_hash_chain);
+
+ if (lh->lh_cookie == cookie)
+ return (lh);
+ }
+
+ return (NULL);
+}
+
+void
+lnet_initialise_handle (lnet_libhandle_t *lh, int type)
{
+ /* ALWAYS called with LNET_LOCK held */
+ unsigned int hash;
+
+ LASSERT (type >= 0 && type < LNET_COOKIE_TYPES);
+ lh->lh_cookie = the_lnet.ln_next_object_cookie | type;
+ the_lnet.ln_next_object_cookie += LNET_COOKIE_TYPES;
+
+ hash = ((unsigned int)lh->lh_cookie) % the_lnet.ln_lh_hash_size;
+ list_add (&lh->lh_hash_chain, &the_lnet.ln_lh_hash_table[hash]);
}
-static void ptl_ni_init_mutex_exit (void)
+void
+lnet_invalidate_handle (lnet_libhandle_t *lh)
{
+ /* ALWAYS called with LNET_LOCK held */
+ list_del (&lh->lh_hash_chain);
+}
+
+int
+lnet_init_finalizers(void)
+{
+#ifdef __KERNEL__
+ int i;
+
+ the_lnet.ln_nfinalizers = num_online_cpus();
+
+ LIBCFS_ALLOC(the_lnet.ln_finalizers,
+ the_lnet.ln_nfinalizers *
+ sizeof(*the_lnet.ln_finalizers));
+ if (the_lnet.ln_finalizers == NULL) {
+ CERROR("Can't allocate ln_finalizers\n");
+ return -ENOMEM;
+ }
+
+ for (i = 0; i < the_lnet.ln_nfinalizers; i++)
+ the_lnet.ln_finalizers[i] = NULL;
+#else
+ the_lnet.ln_finalizing = 0;
+#endif
+
+ CFS_INIT_LIST_HEAD(&the_lnet.ln_finalizeq);
+ return 0;
+}
+
+void
+lnet_fini_finalizers(void)
+{
+#ifdef __KERNEL__
+ int i;
+
+ for (i = 0; i < the_lnet.ln_nfinalizers; i++)
+ LASSERT (the_lnet.ln_finalizers[i] == NULL);
+
+ LIBCFS_FREE(the_lnet.ln_finalizers,
+ the_lnet.ln_nfinalizers *
+ sizeof(*the_lnet.ln_finalizers));
+#else
+ LASSERT (!the_lnet.ln_finalizing);
+#endif
+ LASSERT (list_empty(&the_lnet.ln_finalizeq));
+}
+
+#ifndef __KERNEL__
+/* Temporary workaround to allow uOSS and test programs force server
+ * mode in userspace. See comments near ln_server_mode_flag in
+ * lnet/lib-types.h */
+
+void
+lnet_server_mode() {
+ the_lnet.ln_server_mode_flag = 1;
}
+#endif
+int
+lnet_prepare(lnet_pid_t requested_pid)
+{
+ /* Prepare to bring up the network */
+ int rc = 0;
+ int i;
+
+ LASSERT (the_lnet.ln_refcount == 0);
+
+ the_lnet.ln_routing = 0;
+
+#ifdef __KERNEL__
+ LASSERT ((requested_pid & LNET_PID_USERFLAG) == 0);
+ the_lnet.ln_pid = requested_pid;
+#else
+ if (the_lnet.ln_server_mode_flag) {/* server case (uOSS) */
+ LASSERT ((requested_pid & LNET_PID_USERFLAG) == 0);
+
+ if (cfs_curproc_uid())/* Only root can run user-space server */
+ return -EPERM;
+ the_lnet.ln_pid = requested_pid;
+
+ } else {/* client case (liblustre) */
+
+ /* My PID must be unique on this node and flag I'm userspace */
+ the_lnet.ln_pid = getpid() | LNET_PID_USERFLAG;
+ }
#endif
-int PtlNIInit(ptl_interface_t interface, ptl_pt_index_t ptl_size,
- ptl_ac_index_t acl_size, ptl_pid_t requested_pid,
- ptl_handle_ni_t * handle)
+ rc = lnet_descriptor_setup();
+ if (rc != 0)
+ goto failed0;
+
+ memset(&the_lnet.ln_counters, 0,
+ sizeof(the_lnet.ln_counters));
+
+ CFS_INIT_LIST_HEAD (&the_lnet.ln_active_msgs);
+ CFS_INIT_LIST_HEAD (&the_lnet.ln_active_mds);
+ CFS_INIT_LIST_HEAD (&the_lnet.ln_active_eqs);
+ CFS_INIT_LIST_HEAD (&the_lnet.ln_test_peers);
+ CFS_INIT_LIST_HEAD (&the_lnet.ln_nis);
+ CFS_INIT_LIST_HEAD (&the_lnet.ln_zombie_nis);
+ CFS_INIT_LIST_HEAD (&the_lnet.ln_remote_nets);
+ CFS_INIT_LIST_HEAD (&the_lnet.ln_routers);
+
+ the_lnet.ln_interface_cookie = lnet_create_interface_cookie();
+
+ lnet_init_rtrpools();
+
+ rc = lnet_setup_handle_hash ();
+ if (rc != 0)
+ goto failed0;
+
+ rc = lnet_create_peer_table();
+ if (rc != 0)
+ goto failed1;
+
+ rc = lnet_init_finalizers();
+ if (rc != 0)
+ goto failed2;
+
+ the_lnet.ln_nportals = MAX_PORTALS;
+ LIBCFS_ALLOC(the_lnet.ln_portals,
+ the_lnet.ln_nportals *
+ sizeof(*the_lnet.ln_portals));
+ if (the_lnet.ln_portals == NULL) {
+ rc = -ENOMEM;
+ goto failed3;
+ }
+
+ for (i = 0; i < the_lnet.ln_nportals; i++) {
+ CFS_INIT_LIST_HEAD(&(the_lnet.ln_portals[i].ptl_ml));
+ CFS_INIT_LIST_HEAD(&(the_lnet.ln_portals[i].ptl_msgq));
+ the_lnet.ln_portals[i].ptl_options = 0;
+ }
+
+ return 0;
+
+ failed3:
+ lnet_fini_finalizers();
+ failed2:
+ lnet_destroy_peer_table();
+ failed1:
+ lnet_cleanup_handle_hash();
+ failed0:
+ lnet_descriptor_cleanup();
+ return rc;
+}
+
+int
+lnet_unprepare (void)
+{
+ int idx;
+
+ /* NB no LNET_LOCK since this is the last reference. All LND instances
+ * have shut down already, so it is safe to unlink and free all
+ * descriptors, even those that appear committed to a network op (eg MD
+ * with non-zero pending count) */
+
+ lnet_fail_nid(LNET_NID_ANY, 0);
+
+ LASSERT (list_empty(&the_lnet.ln_test_peers));
+ LASSERT (the_lnet.ln_refcount == 0);
+ LASSERT (list_empty(&the_lnet.ln_nis));
+ LASSERT (list_empty(&the_lnet.ln_zombie_nis));
+ LASSERT (the_lnet.ln_nzombie_nis == 0);
+
+ for (idx = 0; idx < the_lnet.ln_nportals; idx++) {
+ LASSERT (list_empty(&the_lnet.ln_portals[idx].ptl_msgq));
+
+ while (!list_empty (&the_lnet.ln_portals[idx].ptl_ml)) {
+ lnet_me_t *me = list_entry (the_lnet.ln_portals[idx].ptl_ml.next,
+ lnet_me_t, me_list);
+
+ CERROR ("Active me %p on exit\n", me);
+ list_del (&me->me_list);
+ lnet_me_free (me);
+ }
+ }
+
+ while (!list_empty (&the_lnet.ln_active_mds)) {
+ lnet_libmd_t *md = list_entry (the_lnet.ln_active_mds.next,
+ lnet_libmd_t, md_list);
+
+ CERROR ("Active md %p on exit\n", md);
+ list_del (&md->md_list);
+ lnet_md_free (md);
+ }
+
+ while (!list_empty (&the_lnet.ln_active_eqs)) {
+ lnet_eq_t *eq = list_entry (the_lnet.ln_active_eqs.next,
+ lnet_eq_t, eq_list);
+
+ CERROR ("Active eq %p on exit\n", eq);
+ list_del (&eq->eq_list);
+ lnet_eq_free (eq);
+ }
+
+ while (!list_empty (&the_lnet.ln_active_msgs)) {
+ lnet_msg_t *msg = list_entry (the_lnet.ln_active_msgs.next,
+ lnet_msg_t, msg_activelist);
+
+ CERROR ("Active msg %p on exit\n", msg);
+ LASSERT (msg->msg_onactivelist);
+ msg->msg_onactivelist = 0;
+ list_del (&msg->msg_activelist);
+ lnet_msg_free (msg);
+ }
+
+ LIBCFS_FREE(the_lnet.ln_portals,
+ the_lnet.ln_nportals * sizeof(*the_lnet.ln_portals));
+
+ lnet_free_rtrpools();
+ lnet_fini_finalizers();
+ lnet_destroy_peer_table();
+ lnet_cleanup_handle_hash();
+ lnet_descriptor_cleanup();
+
+ return (0);
+}
+
+lnet_ni_t *
+lnet_net2ni_locked (__u32 net)
+{
+ struct list_head *tmp;
+ lnet_ni_t *ni;
+
+ list_for_each (tmp, &the_lnet.ln_nis) {
+ ni = list_entry(tmp, lnet_ni_t, ni_list);
+
+ if (lnet_ptlcompat_matchnet(LNET_NIDNET(ni->ni_nid), net)) {
+ lnet_ni_addref_locked(ni);
+ return ni;
+ }
+ }
+
+ return NULL;
+}
+
+int
+lnet_islocalnet (__u32 net)
+{
+ lnet_ni_t *ni;
+
+ LNET_LOCK();
+ ni = lnet_net2ni_locked(net);
+ if (ni != NULL)
+ lnet_ni_decref_locked(ni);
+ LNET_UNLOCK();
+
+ return ni != NULL;
+}
+
+lnet_ni_t *
+lnet_nid2ni_locked (lnet_nid_t nid)
{
- nal_t *nal;
- int i;
+ struct list_head *tmp;
+ lnet_ni_t *ni;
+
+ list_for_each (tmp, &the_lnet.ln_nis) {
+ ni = list_entry(tmp, lnet_ni_t, ni_list);
+
+ if (lnet_ptlcompat_matchnid(ni->ni_nid, nid)) {
+ lnet_ni_addref_locked(ni);
+ return ni;
+ }
+ }
+
+ return NULL;
+}
+
+int
+lnet_islocalnid (lnet_nid_t nid)
+{
+ lnet_ni_t *ni;
+
+ LNET_LOCK();
+ ni = lnet_nid2ni_locked(nid);
+ if (ni != NULL)
+ lnet_ni_decref_locked(ni);
+ LNET_UNLOCK();
+
+ return ni != NULL;
+}
+
+int
+lnet_count_acceptor_nis (lnet_ni_t **first_ni)
+{
+ /* Return the # of NIs that need the acceptor. Return the first one in
+ * *first_ni so the acceptor can pass it connections "blind" to retain
+ * binary compatibility. */
+ int count = 0;
+#if defined(__KERNEL__) || defined(HAVE_LIBPTHREAD)
+ struct list_head *tmp;
+ lnet_ni_t *ni;
+
+ LNET_LOCK();
+ list_for_each (tmp, &the_lnet.ln_nis) {
+ ni = list_entry(tmp, lnet_ni_t, ni_list);
+
+ if (ni->ni_lnd->lnd_accept != NULL) {
+ /* This LND uses the acceptor */
+ if (count == 0 && first_ni != NULL) {
+ lnet_ni_addref_locked(ni);
+ *first_ni = ni;
+ }
+ count++;
+ }
+ }
+
+ LNET_UNLOCK();
+
+#endif /* defined(__KERNEL__) || defined(HAVE_LIBPTHREAD) */
+ return count;
+}
+
+void
+lnet_shutdown_lndnis (void)
+{
+ int i;
+ int islo;
+ lnet_ni_t *ni;
+
+ /* NB called holding the global mutex */
+
+ /* All quiet on the API front */
+ LASSERT (!the_lnet.ln_shutdown);
+ LASSERT (the_lnet.ln_refcount == 0);
+ LASSERT (list_empty(&the_lnet.ln_zombie_nis));
+ LASSERT (the_lnet.ln_nzombie_nis == 0);
+ LASSERT (list_empty(&the_lnet.ln_remote_nets));
+
+ LNET_LOCK();
+ the_lnet.ln_shutdown = 1; /* flag shutdown */
+
+ /* Unlink NIs from the global table */
+ while (!list_empty(&the_lnet.ln_nis)) {
+ ni = list_entry(the_lnet.ln_nis.next,
+ lnet_ni_t, ni_list);
+ list_del (&ni->ni_list);
- if (!ptl_init)
- return PTL_NOINIT;
+ the_lnet.ln_nzombie_nis++;
+ lnet_ni_decref_locked(ni); /* drop apini's ref */
+ }
+
+ /* Drop the cached eqwait NI. */
+ if (the_lnet.ln_eqwaitni != NULL) {
+ lnet_ni_decref_locked(the_lnet.ln_eqwaitni);
+ the_lnet.ln_eqwaitni = NULL;
+ }
+
+ /* Drop the cached loopback NI. */
+ if (the_lnet.ln_loni != NULL) {
+ lnet_ni_decref_locked(the_lnet.ln_loni);
+ the_lnet.ln_loni = NULL;
+ }
+
+ LNET_UNLOCK();
+
+ /* Clear lazy portals and drop delayed messages which hold refs
+ * on their lnet_msg_t::msg_rxpeer */
+ for (i = 0; i < the_lnet.ln_nportals; i++)
+ LNetClearLazyPortal(i);
+
+ /* Clear the peer table and wait for all peers to go (they hold refs on
+ * their NIs) */
+ lnet_clear_peer_table();
+
+ LNET_LOCK();
+ /* Now wait for the NI's I just nuked to show up on apini_zombie_nis
+ * and shut them down in guaranteed thread context */
+ i = 2;
+ while (the_lnet.ln_nzombie_nis != 0) {
+
+ while (list_empty(&the_lnet.ln_zombie_nis)) {
+ LNET_UNLOCK();
+ ++i;
+ if ((i & (-i)) == i)
+ CDEBUG(D_WARNING,"Waiting for %d zombie NIs\n",
+ the_lnet.ln_nzombie_nis);
+ cfs_pause(cfs_time_seconds(1));
+ LNET_LOCK();
+ }
+
+ ni = list_entry(the_lnet.ln_zombie_nis.next,
+ lnet_ni_t, ni_list);
+ list_del(&ni->ni_list);
+ ni->ni_lnd->lnd_refcount--;
+
+ LNET_UNLOCK();
+
+ islo = ni->ni_lnd->lnd_type == LOLND;
+
+ LASSERT (!in_interrupt ());
+ (ni->ni_lnd->lnd_shutdown)(ni);
- ptl_ni_init_mutex_enter ();
+ /* can't deref lnd anymore now; it might have unregistered
+ * itself... */
- nal = interface(ptl_num_interfaces, ptl_size, acl_size, requested_pid);
+ if (!islo)
+ CDEBUG(D_LNI, "Removed LNI %s\n",
+ libcfs_nid2str(ni->ni_nid));
- if (!nal) {
- ptl_ni_init_mutex_exit ();
- return PTL_NAL_FAILED;
+ LIBCFS_FREE(ni, sizeof(*ni));
+
+ LNET_LOCK();
+ the_lnet.ln_nzombie_nis--;
+ }
+
+ the_lnet.ln_shutdown = 0;
+ LNET_UNLOCK();
+
+ if (the_lnet.ln_network_tokens != NULL) {
+ LIBCFS_FREE(the_lnet.ln_network_tokens,
+ the_lnet.ln_network_tokens_nob);
+ the_lnet.ln_network_tokens = NULL;
}
+}
+
+int
+lnet_startup_lndnis (void)
+{
+ lnd_t *lnd;
+ lnet_ni_t *ni;
+ struct list_head nilist;
+ int rc = 0;
+ int lnd_type;
+ int nicount = 0;
+ char *nets = lnet_get_networks();
+
+ CFS_INIT_LIST_HEAD(&nilist);
+
+ if (nets == NULL)
+ goto failed;
+
+ rc = lnet_parse_networks(&nilist, nets);
+ if (rc != 0)
+ goto failed;
+
+ while (!list_empty(&nilist)) {
+ ni = list_entry(nilist.next, lnet_ni_t, ni_list);
+ lnd_type = LNET_NETTYP(LNET_NIDNET(ni->ni_nid));
+
+ LASSERT (libcfs_isknown_lnd(lnd_type));
+
+ LNET_MUTEX_DOWN(&the_lnet.ln_lnd_mutex);
+ lnd = lnet_find_lnd_by_type(lnd_type);
+
+#ifdef __KERNEL__
+ if (lnd == NULL) {
+ LNET_MUTEX_UP(&the_lnet.ln_lnd_mutex);
+ rc = request_module(libcfs_lnd2modname(lnd_type));
+ LNET_MUTEX_DOWN(&the_lnet.ln_lnd_mutex);
- for (i = 0; i < ptl_num_interfaces; i++) {
- if (ptl_interfaces[i] == nal) {
- nal->refct++;
- handle->nal_idx = (NI_HANDLE_MAGIC & ~NI_HANDLE_MASK) | i;
- fprintf(stderr, "Returning existing NAL (%d)\n", i);
- ptl_ni_init_mutex_exit ();
- return PTL_OK;
+ lnd = lnet_find_lnd_by_type(lnd_type);
+ if (lnd == NULL) {
+ LNET_MUTEX_UP(&the_lnet.ln_lnd_mutex);
+ CERROR("Can't load LND %s, module %s, rc=%d\n",
+ libcfs_lnd2str(lnd_type),
+ libcfs_lnd2modname(lnd_type), rc);
+#ifndef CONFIG_KMOD
+ LCONSOLE_ERROR_MSG(0x104, "Your kernel must be "
+ "compiled with CONFIG_KMOD set for "
+ "automatic module loading.");
+#endif
+ goto failed;
+ }
+ }
+#else
+ if (lnd == NULL) {
+ LNET_MUTEX_UP(&the_lnet.ln_lnd_mutex);
+ CERROR("LND %s not supported\n",
+ libcfs_lnd2str(lnd_type));
+ goto failed;
}
+#endif
+
+ ni->ni_refcount = 1;
+
+ LNET_LOCK();
+ lnd->lnd_refcount++;
+ LNET_UNLOCK();
+
+ ni->ni_lnd = lnd;
+
+ rc = (lnd->lnd_startup)(ni);
+
+ LNET_MUTEX_UP(&the_lnet.ln_lnd_mutex);
+
+ if (rc != 0) {
+ LCONSOLE_ERROR_MSG(0x105, "Error %d starting up LNI %s"
+ "\n",
+ rc, libcfs_lnd2str(lnd->lnd_type));
+ LNET_LOCK();
+ lnd->lnd_refcount--;
+ LNET_UNLOCK();
+ goto failed;
+ }
+
+ list_del(&ni->ni_list);
+
+ LNET_LOCK();
+ list_add_tail(&ni->ni_list, &the_lnet.ln_nis);
+ LNET_UNLOCK();
+
+ if (lnd->lnd_type == LOLND) {
+ lnet_ni_addref(ni);
+ LASSERT (the_lnet.ln_loni == NULL);
+ the_lnet.ln_loni = ni;
+ continue;
+ }
+
+#ifndef __KERNEL__
+ if (lnd->lnd_wait != NULL) {
+ if (the_lnet.ln_eqwaitni == NULL) {
+ lnet_ni_addref(ni);
+ the_lnet.ln_eqwaitni = ni;
+ }
+ } else {
+# ifndef HAVE_LIBPTHREAD
+ LCONSOLE_ERROR_MSG(0x106, "LND %s not supported in a "
+ "single-threaded runtime\n",
+ libcfs_lnd2str(lnd_type));
+ goto failed;
+# endif
+ }
+#endif
+ if (ni->ni_peertxcredits == 0 ||
+ ni->ni_maxtxcredits == 0) {
+ LCONSOLE_ERROR_MSG(0x107, "LNI %s has no %scredits\n",
+ libcfs_lnd2str(lnd->lnd_type),
+ ni->ni_peertxcredits == 0 ?
+ "" : "per-peer ");
+ goto failed;
+ }
+
+ ni->ni_txcredits = ni->ni_mintxcredits = ni->ni_maxtxcredits;
+
+ CDEBUG(D_LNI, "Added LNI %s [%d/%d]\n",
+ libcfs_nid2str(ni->ni_nid),
+ ni->ni_peertxcredits, ni->ni_txcredits);
+
+ /* Handle nidstrings for network 0 just like this one */
+ if (the_lnet.ln_ptlcompat > 0) {
+ if (nicount > 0) {
+ LCONSOLE_ERROR_MSG(0x108, "Can't run > 1 "
+ "network when portals_compatibility is "
+ "set\n");
+ goto failed;
+ }
+ libcfs_setnet0alias(lnd->lnd_type);
+ }
+
+ nicount++;
+ }
+
+ if (the_lnet.ln_eqwaitni != NULL && nicount > 1) {
+ lnd_type = the_lnet.ln_eqwaitni->ni_lnd->lnd_type;
+ LCONSOLE_ERROR_MSG(0x109, "LND %s can only run single-network"
+ "\n",
+ libcfs_lnd2str(lnd_type));
+ goto failed;
+ }
+
+ return 0;
+
+ failed:
+ lnet_shutdown_lndnis();
+
+ while (!list_empty(&nilist)) {
+ ni = list_entry(nilist.next, lnet_ni_t, ni_list);
+ list_del(&ni->ni_list);
+ LIBCFS_FREE(ni, sizeof(*ni));
}
- nal->refct = 1;
- if (ptl_num_interfaces >= MAX_NIS) {
- if (nal->shutdown)
- nal->shutdown (nal, ptl_num_interfaces);
- ptl_ni_init_mutex_exit ();
- return PTL_NOSPACE;
+ return -ENETDOWN;
+}
+
+int
+LNetInit(void)
+{
+ int rc;
+
+ lnet_assert_wire_constants ();
+ LASSERT (!the_lnet.ln_init);
+
+ memset(&the_lnet, 0, sizeof(the_lnet));
+
+ rc = lnet_get_portals_compatibility();
+ if (rc < 0)
+ return rc;
+
+ lnet_init_locks();
+ CFS_INIT_LIST_HEAD(&the_lnet.ln_lnds);
+ the_lnet.ln_ptlcompat = rc;
+ the_lnet.ln_refcount = 0;
+ the_lnet.ln_init = 1;
+
+#ifdef __KERNEL__
+ /* All LNDs apart from the LOLND are in separate modules. They
+ * register themselves when their module loads, and unregister
+ * themselves when their module is unloaded. */
+#else
+ /* Register LNDs
+ * NB the order here determines default 'networks=' order */
+# ifdef CRAY_XT3
+ LNET_REGISTER_ULND(the_ptllnd);
+# endif
+# ifdef HAVE_LIBPTHREAD
+ LNET_REGISTER_ULND(the_tcplnd);
+# endif
+#endif
+ lnet_register_lnd(&the_lolnd);
+ return 0;
+}
+
+void
+LNetFini(void)
+{
+ LASSERT (the_lnet.ln_init);
+ LASSERT (the_lnet.ln_refcount == 0);
+
+ while (!list_empty(&the_lnet.ln_lnds))
+ lnet_unregister_lnd(list_entry(the_lnet.ln_lnds.next,
+ lnd_t, lnd_list));
+ lnet_fini_locks();
+
+ the_lnet.ln_init = 0;
+}
+
+int
+LNetNIInit(lnet_pid_t requested_pid)
+{
+ int im_a_router = 0;
+ int rc;
+
+ LNET_MUTEX_DOWN(&the_lnet.ln_api_mutex);
+
+ LASSERT (the_lnet.ln_init);
+ CDEBUG(D_OTHER, "refs %d\n", the_lnet.ln_refcount);
+
+ if (the_lnet.ln_refcount > 0) {
+ rc = the_lnet.ln_refcount++;
+ goto out;
+ }
+
+ if (requested_pid == LNET_PID_ANY) {
+ /* Don't instantiate LNET just for me */
+ rc = -ENETDOWN;
+ goto failed0;
+ }
+
+ rc = lnet_prepare(requested_pid);
+ if (rc != 0)
+ goto failed0;
+
+ rc = lnet_startup_lndnis();
+ if (rc != 0)
+ goto failed1;
+
+ rc = lnet_parse_routes(lnet_get_routes(), &im_a_router);
+ if (rc != 0)
+ goto failed2;
+
+ rc = lnet_check_routes();
+ if (rc != 0)
+ goto failed2;
+
+ rc = lnet_alloc_rtrpools(im_a_router);
+ if (rc != 0)
+ goto failed2;
+
+ rc = lnet_acceptor_start();
+ if (rc != 0)
+ goto failed2;
+
+ the_lnet.ln_refcount = 1;
+ /* Now I may use my own API functions... */
+
+ rc = lnet_router_checker_start();
+ if (rc != 0)
+ goto failed3;
+
+ rc = lnet_ping_target_init();
+ if (rc != 0)
+ goto failed4;
+
+ lnet_proc_init();
+ goto out;
+
+ failed4:
+ lnet_router_checker_stop();
+ failed3:
+ the_lnet.ln_refcount = 0;
+ lnet_acceptor_stop();
+ failed2:
+ lnet_destroy_routes();
+ lnet_shutdown_lndnis();
+ failed1:
+ lnet_unprepare();
+ failed0:
+ LASSERT (rc < 0);
+ out:
+ LNET_MUTEX_UP(&the_lnet.ln_api_mutex);
+ return rc;
+}
+
+int
+LNetNIFini()
+{
+ LNET_MUTEX_DOWN(&the_lnet.ln_api_mutex);
+
+ LASSERT (the_lnet.ln_init);
+ LASSERT (the_lnet.ln_refcount > 0);
+
+ if (the_lnet.ln_refcount != 1) {
+ the_lnet.ln_refcount--;
+ } else {
+ LASSERT (!the_lnet.ln_niinit_self);
+
+ lnet_proc_fini();
+ lnet_ping_target_fini();
+ lnet_router_checker_stop();
+
+ /* Teardown fns that use my own API functions BEFORE here */
+ the_lnet.ln_refcount = 0;
+
+ lnet_acceptor_stop();
+ lnet_destroy_routes();
+ lnet_shutdown_lndnis();
+ lnet_unprepare();
}
- handle->nal_idx = (NI_HANDLE_MAGIC & ~NI_HANDLE_MASK) | ptl_num_interfaces;
- ptl_interfaces[ptl_num_interfaces++] = nal;
+ LNET_MUTEX_UP(&the_lnet.ln_api_mutex);
+ return 0;
+}
+
+int
+LNetCtl(unsigned int cmd, void *arg)
+{
+ struct libcfs_ioctl_data *data = arg;
+ lnet_process_id_t id;
+ lnet_ni_t *ni;
+ int rc;
+
+ LASSERT (the_lnet.ln_init);
+ LASSERT (the_lnet.ln_refcount > 0);
+
+ switch (cmd) {
+ case IOC_LIBCFS_GET_NI:
+ rc = LNetGetId(data->ioc_count, &id);
+ data->ioc_nid = id.nid;
+ return rc;
+
+ case IOC_LIBCFS_FAIL_NID:
+ return lnet_fail_nid(data->ioc_nid, data->ioc_count);
+
+ case IOC_LIBCFS_ADD_ROUTE:
+ rc = lnet_add_route(data->ioc_net, data->ioc_count,
+ data->ioc_nid);
+ return (rc != 0) ? rc : lnet_check_routes();
+
+ case IOC_LIBCFS_DEL_ROUTE:
+ return lnet_del_route(data->ioc_net, data->ioc_nid);
+
+ case IOC_LIBCFS_GET_ROUTE:
+ return lnet_get_route(data->ioc_count,
+ &data->ioc_net, &data->ioc_count,
+ &data->ioc_nid, &data->ioc_flags);
+ case IOC_LIBCFS_NOTIFY_ROUTER:
+ return lnet_notify(NULL, data->ioc_nid, data->ioc_flags,
+ (time_t)data->ioc_u64[0]);
+
+ case IOC_LIBCFS_PORTALS_COMPATIBILITY:
+ return the_lnet.ln_ptlcompat;
+
+ case IOC_LIBCFS_LNET_DIST:
+ rc = LNetDist(data->ioc_nid, &data->ioc_nid, &data->ioc_u32[1]);
+ if (rc < 0 && rc != -EHOSTUNREACH)
+ return rc;
+
+ data->ioc_u32[0] = rc;
+ return 0;
+
+ case IOC_LIBCFS_TESTPROTOCOMPAT:
+ LNET_LOCK();
+ the_lnet.ln_testprotocompat = data->ioc_flags;
+ LNET_UNLOCK();
+ return 0;
+
+ case IOC_LIBCFS_PING:
+ rc = lnet_ping((lnet_process_id_t) {.nid = data->ioc_nid,
+ .pid = data->ioc_u32[0]},
+ data->ioc_u32[1], /* timeout */
+ (lnet_process_id_t *)data->ioc_pbuf1,
+ data->ioc_plen1/sizeof(lnet_process_id_t));
+ if (rc < 0)
+ return rc;
+ data->ioc_count = rc;
+ return 0;
- ptl_eq_ni_init(nal);
- ptl_me_ni_init(nal);
+ case IOC_LIBCFS_DEBUG_PEER: {
+ /* CAVEAT EMPTOR: this one designed for calling directly; not
+ * via an ioctl */
+ lnet_process_id_t *id = arg;
- ptl_ni_init_mutex_exit ();
- return PTL_OK;
+ lnet_debug_peer(id->nid);
+
+ ni = lnet_net2ni(LNET_NIDNET(id->nid));
+ if (ni == NULL) {
+ CDEBUG(D_WARNING, "No NI for %s\n", libcfs_id2str(*id));
+ } else {
+ if (ni->ni_lnd->lnd_ctl == NULL) {
+ CDEBUG(D_WARNING, "No ctl for %s\n",
+ libcfs_id2str(*id));
+ } else {
+ (void)ni->ni_lnd->lnd_ctl(ni, cmd, arg);
+ }
+
+ lnet_ni_decref(ni);
+ }
+ return 0;
+ }
+
+ default:
+ ni = lnet_net2ni(data->ioc_net);
+ if (ni == NULL)
+ return -EINVAL;
+
+ if (ni->ni_lnd->lnd_ctl == NULL)
+ rc = -EINVAL;
+ else
+ rc = ni->ni_lnd->lnd_ctl(ni, cmd, arg);
+
+ lnet_ni_decref(ni);
+ return rc;
+ }
+ /* not reached */
}
+int
+LNetGetId(unsigned int index, lnet_process_id_t *id)
+{
+ lnet_ni_t *ni;
+ struct list_head *tmp;
+ int rc = -ENOENT;
+
+ LASSERT (the_lnet.ln_init);
+ LASSERT (the_lnet.ln_refcount > 0);
-int PtlNIFini(ptl_handle_ni_t ni)
+ LNET_LOCK();
+
+ list_for_each(tmp, &the_lnet.ln_nis) {
+ if (index-- != 0)
+ continue;
+
+ ni = list_entry(tmp, lnet_ni_t, ni_list);
+
+ id->nid = ni->ni_nid;
+ id->pid = the_lnet.ln_pid;
+ rc = 0;
+ break;
+ }
+
+ LNET_UNLOCK();
+
+ return rc;
+}
+
+void
+LNetSnprintHandle(char *str, int len, lnet_handle_any_t h)
{
- nal_t *nal;
- int idx;
- int rc;
+ snprintf(str, len, LPX64, h.cookie);
+}
- if (!ptl_init)
- return PTL_NOINIT;
- ptl_ni_init_mutex_enter ();
+int
+lnet_ping_target_init(void)
+{
+ lnet_handle_me_t meh;
+ lnet_process_id_t id;
+ int rc;
+ int rc2;
+ int n;
+ int infosz;
+ int i;
+
+ for (n = 0; ; n++) {
+ rc = LNetGetId(n, &id);
+ if (rc == -ENOENT)
+ break;
+
+ LASSERT (rc == 0);
+ }
- nal = ptl_hndl2nal (&ni);
- if (nal == NULL) {
- ptl_ni_init_mutex_exit ();
- return PTL_INV_HANDLE;
+ infosz = offsetof(lnet_ping_info_t, pi_nid[n]);
+ LIBCFS_ALLOC(the_lnet.ln_ping_info, infosz);
+ if (the_lnet.ln_ping_info == NULL) {
+ CERROR("Can't allocate ping info[%d]\n", n);
+ return -ENOMEM;
}
- idx = ni.nal_idx & NI_HANDLE_MASK;
+ the_lnet.ln_ping_info->pi_magic = LNET_PROTO_PING_MAGIC;
+ the_lnet.ln_ping_info->pi_version = LNET_PROTO_PING_VERSION;
+ the_lnet.ln_ping_info->pi_pid = the_lnet.ln_pid;
+ the_lnet.ln_ping_info->pi_nnids = n;
- nal->refct--;
- if (nal->refct > 0) {
- ptl_ni_init_mutex_exit ();
- return PTL_OK;
+ for (i = 0; i < n; i++) {
+ rc = LNetGetId(i, &id);
+ LASSERT (rc == 0);
+ the_lnet.ln_ping_info->pi_nid[i] = id.nid;
+ }
+
+ /* We can have a tiny EQ since we only need to see the unlink event on
+ * teardown, which by definition is the last one! */
+ rc = LNetEQAlloc(2, LNET_EQ_HANDLER_NONE, &the_lnet.ln_ping_target_eq);
+ if (rc != 0) {
+ CERROR("Can't allocate ping EQ: %d\n", rc);
+ goto failed_0;
}
- ptl_me_ni_fini(nal);
- ptl_eq_ni_fini(nal);
+ rc = LNetMEAttach(LNET_RESERVED_PORTAL,
+ (lnet_process_id_t){.nid = LNET_NID_ANY,
+ .pid = LNET_PID_ANY},
+ LNET_PROTO_PING_MATCHBITS, 0LL,
+ LNET_UNLINK, LNET_INS_AFTER,
+ &meh);
+ if (rc != 0) {
+ CERROR("Can't create ping ME: %d\n", rc);
+ goto failed_1;
+ }
- rc = PTL_OK;
- if (nal->shutdown)
- rc = nal->shutdown(nal, idx);
+ rc = LNetMDAttach(meh,
+ (lnet_md_t){.start = the_lnet.ln_ping_info,
+ .length = infosz,
+ .threshold = LNET_MD_THRESH_INF,
+ .options = (LNET_MD_OP_GET |
+ LNET_MD_TRUNCATE |
+ LNET_MD_MANAGE_REMOTE),
+ .eq_handle = the_lnet.ln_ping_target_eq},
+ LNET_RETAIN,
+ &the_lnet.ln_ping_target_md);
+ if (rc != 0) {
+ CERROR("Can't attach ping MD: %d\n", rc);
+ goto failed_2;
+ }
- ptl_interfaces[idx] = NULL;
- ptl_num_interfaces--;
+ return 0;
+
+ failed_2:
+ rc2 = LNetMEUnlink(meh);
+ LASSERT (rc2 == 0);
+ failed_1:
+ rc2 = LNetEQFree(the_lnet.ln_ping_target_eq);
+ LASSERT (rc2 == 0);
+ failed_0:
+ LIBCFS_FREE(the_lnet.ln_ping_info, infosz);
- ptl_ni_init_mutex_exit ();
return rc;
}
-int PtlNIHandle(ptl_handle_any_t handle_in, ptl_handle_ni_t * ni_out)
+void
+lnet_ping_target_fini(void)
+{
+ lnet_event_t event;
+ int rc;
+ int which;
+ int timeout_ms = 1000;
+ cfs_sigset_t blocked = cfs_block_allsigs();
+
+ LNetMDUnlink(the_lnet.ln_ping_target_md);
+ /* NB md could be busy; this just starts the unlink */
+
+ for (;;) {
+ rc = LNetEQPoll(&the_lnet.ln_ping_target_eq, 1,
+ timeout_ms, &event, &which);
+
+ /* I expect overflow... */
+ LASSERT (rc >= 0 || rc == -EOVERFLOW);
+
+ if (rc == 0) {
+ /* timed out: provide a diagnostic */
+ CWARN("Still waiting for ping MD to unlink\n");
+ timeout_ms *= 2;
+ continue;
+ }
+
+ /* Got a valid event */
+ if (event.unlinked)
+ break;
+ }
+
+ rc = LNetEQFree(the_lnet.ln_ping_target_eq);
+ LASSERT (rc == 0);
+
+ LIBCFS_FREE(the_lnet.ln_ping_info,
+ offsetof(lnet_ping_info_t,
+ pi_nid[the_lnet.ln_ping_info->pi_nnids]));
+
+ cfs_restore_sigs(blocked);
+}
+
+int
+lnet_ping (lnet_process_id_t id, int timeout_ms, lnet_process_id_t *ids, int n_ids)
{
- *ni_out = handle_in;
+ lnet_handle_eq_t eqh;
+ lnet_handle_md_t mdh;
+ lnet_event_t event;
+ int which;
+ int unlinked = 0;
+ int replied = 0;
+ const int a_long_time = 60000; /* mS */
+ int infosz = offsetof(lnet_ping_info_t, pi_nid[n_ids]);
+ lnet_ping_info_t *info;
+ lnet_process_id_t tmpid;
+ int i;
+ int nob;
+ int rc;
+ int rc2;
+ cfs_sigset_t blocked;
+
+ if (n_ids <= 0 ||
+ id.nid == LNET_NID_ANY ||
+ timeout_ms > 500000 || /* arbitrary limit! */
+ n_ids > 20) /* arbitrary limit! */
+ return -EINVAL;
+
+ if (id.pid == LNET_PID_ANY)
+ id.pid = LUSTRE_SRV_LNET_PID;
+
+ LIBCFS_ALLOC(info, infosz);
+ if (info == NULL)
+ return -ENOMEM;
+
+ /* NB 2 events max (including any unlink event) */
+ rc = LNetEQAlloc(2, LNET_EQ_HANDLER_NONE, &eqh);
+ if (rc != 0) {
+ CERROR("Can't allocate EQ: %d\n", rc);
+ goto out_0;
+ }
+
+ rc = LNetMDBind((lnet_md_t){.start = info,
+ .length = infosz,
+ .threshold = 2, /* GET/REPLY */
+ .options = LNET_MD_TRUNCATE,
+ .eq_handle = eqh},
+ LNET_UNLINK,
+ &mdh);
+ if (rc != 0) {
+ CERROR("Can't bind MD: %d\n", rc);
+ goto out_1;
+ }
+
+ rc = LNetGet(LNET_NID_ANY, mdh, id,
+ LNET_RESERVED_PORTAL,
+ LNET_PROTO_PING_MATCHBITS, 0);
+
+ if (rc != 0) {
+ /* Don't CERROR; this could be deliberate! */
+
+ rc2 = LNetMDUnlink(mdh);
+ LASSERT (rc2 == 0);
+
+ /* NB must wait for the UNLINK event below... */
+ unlinked = 1;
+ timeout_ms = a_long_time;
+ }
+
+ do {
+ /* MUST block for unlink to complete */
+ if (unlinked)
+ blocked = cfs_block_allsigs();
+
+ rc2 = LNetEQPoll(&eqh, 1, timeout_ms, &event, &which);
+
+ if (unlinked)
+ cfs_restore_sigs(blocked);
+
+ CDEBUG(D_NET, "poll %d(%d %d)%s\n", rc2,
+ (rc2 <= 0) ? -1 : event.type,
+ (rc2 <= 0) ? -1 : event.status,
+ (rc2 > 0 && event.unlinked) ? " unlinked" : "");
+
+ LASSERT (rc2 != -EOVERFLOW); /* can't miss anything */
- return PTL_OK;
+ if (rc2 <= 0 || event.status != 0) {
+ /* timeout or error */
+ if (!replied && rc == 0)
+ rc = (rc2 < 0) ? rc2 :
+ (rc2 == 0) ? -ETIMEDOUT :
+ event.status;
+
+ if (!unlinked) {
+ /* Ensure completion in finite time... */
+ LNetMDUnlink(mdh);
+ /* No assertion (racing with network) */
+ unlinked = 1;
+ timeout_ms = a_long_time;
+ } else if (rc2 == 0) {
+ /* timed out waiting for unlink */
+ CWARN("ping %s: late network completion\n",
+ libcfs_id2str(id));
+ }
+
+ } else if (event.type == LNET_EVENT_REPLY) {
+ replied = 1;
+ rc = event.mlength;
+ }
+
+ } while (rc2 <= 0 || !event.unlinked);
+
+ if (!replied) {
+ if (rc >= 0)
+ CWARN("%s: Unexpected rc >= 0 but no reply!\n",
+ libcfs_id2str(id));
+ rc = -EIO;
+ goto out_1;
+ }
+
+ nob = rc;
+ LASSERT (nob >= 0 && nob <= infosz);
+
+ rc = -EPROTO; /* if I can't parse... */
+
+ if (nob < 8) {
+ /* can't check magic/version */
+ CERROR("%s: ping info too short %d\n",
+ libcfs_id2str(id), nob);
+ goto out_1;
+ }
+
+ if (info->pi_magic == __swab32(LNET_PROTO_PING_MAGIC)) {
+ /* NB I might be swabbing garbage until I check below, but it
+ * doesn't matter */
+ __swab32s(&info->pi_version);
+ __swab32s(&info->pi_pid);
+ __swab32s(&info->pi_nnids);
+ for (i = 0; i < info->pi_nnids && i < n_ids; i++)
+ __swab64s(&info->pi_nid[i]);
+
+ } else if (info->pi_magic != LNET_PROTO_PING_MAGIC) {
+ CERROR("%s: Unexpected magic %08x\n",
+ libcfs_id2str(id), info->pi_magic);
+ goto out_1;
+ }
+
+ if (info->pi_version != LNET_PROTO_PING_VERSION) {
+ CERROR("%s: Unexpected version 0x%x\n",
+ libcfs_id2str(id), info->pi_version);
+ goto out_1;
+ }
+
+ if (nob < offsetof(lnet_ping_info_t, pi_nid[0])) {
+ CERROR("%s: Short reply %d(%d min)\n", libcfs_id2str(id),
+ nob, (int)offsetof(lnet_ping_info_t, pi_nid[0]));
+ goto out_1;
+ }
+
+ if (info->pi_nnids < n_ids)
+ n_ids = info->pi_nnids;
+
+ if (nob < offsetof(lnet_ping_info_t, pi_nid[n_ids])) {
+ CERROR("%s: Short reply %d(%d expected)\n", libcfs_id2str(id),
+ nob, (int)offsetof(lnet_ping_info_t, pi_nid[n_ids]));
+ goto out_1;
+ }
+
+ rc = -EFAULT; /* If I SEGV... */
+
+ for (i = 0; i < n_ids; i++) {
+ tmpid.pid = info->pi_pid;
+ tmpid.nid = info->pi_nid[i];
+#ifdef __KERNEL__
+ if (copy_to_user(&ids[i], &tmpid, sizeof(tmpid)))
+ goto out_1;
+#else
+ ids[i] = tmpid;
+#endif
+ }
+ rc = info->pi_nnids;
+
+ out_1:
+ rc2 = LNetEQFree(eqh);
+ if (rc2 != 0)
+ CERROR("rc2 %d\n", rc2);
+ LASSERT (rc2 == 0);
+
+ out_0:
+ LIBCFS_FREE(info, infosz);
+ return rc;
}