X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=blobdiff_plain;f=lnet%2Fklnds%2Fsocklnd%2Fsocklnd.c;h=52492271f83e22109803092e9b76213a920c841e;hp=e3e037a77e53a4c4682340cd07f80a6a017444b6;hb=a7c9aba5eb96dd1e53899108a65af381b49e657b;hpb=fb983bbebf8127828dba545e7b778c12e5411f64 diff --git a/lnet/klnds/socklnd/socklnd.c b/lnet/klnds/socklnd/socklnd.c index e3e037a..5249227 100644 --- a/lnet/klnds/socklnd/socklnd.c +++ b/lnet/klnds/socklnd/socklnd.c @@ -37,10 +37,10 @@ * Author: Eric Barton */ -#include "socklnd.h" #include +#include "socklnd.h" -static struct lnet_lnd the_ksocklnd; +static const struct lnet_lnd the_ksocklnd; struct ksock_nal_data ksocknal_data; static struct ksock_interface * @@ -61,6 +61,58 @@ ksocknal_ip2iface(struct lnet_ni *ni, __u32 ip) return NULL; } +static struct ksock_interface * +ksocknal_index2iface(struct lnet_ni *ni, int index) +{ + struct ksock_net *net = ni->ni_data; + int i; + struct ksock_interface *iface; + + for (i = 0; i < net->ksnn_ninterfaces; i++) { + LASSERT(i < LNET_INTERFACES_NUM); + iface = &net->ksnn_interfaces[i]; + + if (iface->ksni_index == index) + return iface; + } + + return NULL; +} + +static int ksocknal_ip2index(__u32 ipaddress, struct lnet_ni *ni) +{ + struct net_device *dev; + int ret = -1; + DECLARE_CONST_IN_IFADDR(ifa); + + rcu_read_lock(); + for_each_netdev(ni->ni_net_ns, dev) { + int flags = dev_get_flags(dev); + struct in_device *in_dev; + + if (flags & IFF_LOOPBACK) /* skip the loopback IF */ + continue; + + if (!(flags & IFF_UP)) + continue; + + in_dev = __in_dev_get_rcu(dev); + if (!in_dev) + continue; + + in_dev_for_each_ifa_rcu(ifa, in_dev) { + if (ntohl(ifa->ifa_local) == ipaddress) + ret = dev->ifindex; + } + endfor_ifa(in_dev); + if (ret >= 0) + break; + } + rcu_read_unlock(); + + return ret; +} + static struct ksock_route * ksocknal_create_route(__u32 ipaddr, int port) { @@ -74,15 +126,16 @@ ksocknal_create_route(__u32 ipaddr, int port) route->ksnr_peer = NULL; route->ksnr_retry_interval = 0; /* OK to connect at any time */ route->ksnr_ipaddr = ipaddr; - route->ksnr_port = port; - route->ksnr_scheduled = 0; - route->ksnr_connecting = 0; - route->ksnr_connected = 0; - route->ksnr_deleted = 0; - route->ksnr_conn_count = 0; - route->ksnr_share_count = 0; - - return (route); + route->ksnr_myiface = -1; + route->ksnr_port = port; + route->ksnr_scheduled = 0; + route->ksnr_connecting = 0; + route->ksnr_connected = 0; + route->ksnr_deleted = 0; + route->ksnr_conn_count = 0; + route->ksnr_share_count = 0; + + return route; } void @@ -158,19 +211,17 @@ ksocknal_destroy_peer(struct ksock_peer_ni *peer_ni) * state to do with this peer_ni has been cleaned up when its refcount * drops to zero. */ - atomic_dec(&net->ksnn_npeers); + if (atomic_dec_and_test(&net->ksnn_npeers)) + wake_up_var(&net->ksnn_npeers); } struct ksock_peer_ni * ksocknal_find_peer_locked(struct lnet_ni *ni, struct lnet_process_id id) { - struct list_head *peer_list = ksocknal_nid2peerlist(id.nid); - struct list_head *tmp; struct ksock_peer_ni *peer_ni; - list_for_each(tmp, peer_list) { - peer_ni = list_entry(tmp, struct ksock_peer_ni, ksnp_list); - + hash_for_each_possible(ksocknal_data.ksnd_peers, peer_ni, + ksnp_list, id.nid) { LASSERT(!peer_ni->ksnp_closing); if (peer_ni->ksnp_ni != ni) @@ -229,7 +280,7 @@ ksocknal_unlink_peer_locked(struct ksock_peer_ni *peer_ni) LASSERT(list_empty(&peer_ni->ksnp_routes)); LASSERT(!peer_ni->ksnp_closing); peer_ni->ksnp_closing = 1; - list_del(&peer_ni->ksnp_list); + hlist_del(&peer_ni->ksnp_list); /* lose peerlist's ref */ ksocknal_peer_decref(peer_ni); } @@ -240,7 +291,6 @@ ksocknal_get_peer_info(struct lnet_ni *ni, int index, int *port, int *conn_count, int *share_count) { struct ksock_peer_ni *peer_ni; - struct list_head *ptmp; struct ksock_route *route; struct list_head *rtmp; int i; @@ -249,58 +299,56 @@ ksocknal_get_peer_info(struct lnet_ni *ni, int index, read_lock(&ksocknal_data.ksnd_global_lock); - for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++) { - list_for_each(ptmp, &ksocknal_data.ksnd_peers[i]) { - peer_ni = list_entry(ptmp, struct ksock_peer_ni, ksnp_list); - - if (peer_ni->ksnp_ni != ni) - continue; + hash_for_each(ksocknal_data.ksnd_peers, i, peer_ni, ksnp_list) { - if (peer_ni->ksnp_n_passive_ips == 0 && - list_empty(&peer_ni->ksnp_routes)) { - if (index-- > 0) - continue; + if (peer_ni->ksnp_ni != ni) + continue; - *id = peer_ni->ksnp_id; - *myip = 0; - *peer_ip = 0; - *port = 0; - *conn_count = 0; - *share_count = 0; - rc = 0; - goto out; - } + if (peer_ni->ksnp_n_passive_ips == 0 && + list_empty(&peer_ni->ksnp_routes)) { + if (index-- > 0) + continue; - for (j = 0; j < peer_ni->ksnp_n_passive_ips; j++) { - if (index-- > 0) - continue; + *id = peer_ni->ksnp_id; + *myip = 0; + *peer_ip = 0; + *port = 0; + *conn_count = 0; + *share_count = 0; + rc = 0; + goto out; + } - *id = peer_ni->ksnp_id; - *myip = peer_ni->ksnp_passive_ips[j]; - *peer_ip = 0; - *port = 0; - *conn_count = 0; - *share_count = 0; - rc = 0; - goto out; - } + for (j = 0; j < peer_ni->ksnp_n_passive_ips; j++) { + if (index-- > 0) + continue; - list_for_each(rtmp, &peer_ni->ksnp_routes) { - if (index-- > 0) - continue; + *id = peer_ni->ksnp_id; + *myip = peer_ni->ksnp_passive_ips[j]; + *peer_ip = 0; + *port = 0; + *conn_count = 0; + *share_count = 0; + rc = 0; + goto out; + } - route = list_entry(rtmp, struct ksock_route, - ksnr_list); + list_for_each(rtmp, &peer_ni->ksnp_routes) { + if (index-- > 0) + continue; - *id = peer_ni->ksnp_id; - *myip = route->ksnr_myipaddr; - *peer_ip = route->ksnr_ipaddr; - *port = route->ksnr_port; - *conn_count = route->ksnr_conn_count; - *share_count = route->ksnr_share_count; - rc = 0; - goto out; - } + route = list_entry(rtmp, struct ksock_route, + ksnr_list); + + *id = peer_ni->ksnp_id; + rc = choose_ipv4_src(myip, route->ksnr_myiface, + route->ksnr_ipaddr, + ni->ni_net_ns); + *peer_ip = route->ksnr_ipaddr; + *port = route->ksnr_port; + *conn_count = route->ksnr_conn_count; + *share_count = route->ksnr_share_count; + goto out; } } out: @@ -309,47 +357,52 @@ out: } static void -ksocknal_associate_route_conn_locked(struct ksock_route *route, struct ksock_conn *conn) +ksocknal_associate_route_conn_locked(struct ksock_route *route, + struct ksock_conn *conn) { struct ksock_peer_ni *peer_ni = route->ksnr_peer; int type = conn->ksnc_type; struct ksock_interface *iface; + int conn_iface = ksocknal_ip2index(conn->ksnc_myipaddr, + route->ksnr_peer->ksnp_ni); conn->ksnc_route = route; ksocknal_route_addref(route); - if (route->ksnr_myipaddr != conn->ksnc_myipaddr) { - if (route->ksnr_myipaddr == 0) { + if (route->ksnr_myiface != conn_iface) { + if (route->ksnr_myiface < 0) { /* route wasn't bound locally yet (the initial route) */ - CDEBUG(D_NET, "Binding %s %pI4h to %pI4h\n", + CDEBUG(D_NET, "Binding %s %pI4h to interface %d\n", libcfs_id2str(peer_ni->ksnp_id), &route->ksnr_ipaddr, - &conn->ksnc_myipaddr); + conn_iface); } else { - CDEBUG(D_NET, "Rebinding %s %pI4h from %pI4h " - "to %pI4h\n", libcfs_id2str(peer_ni->ksnp_id), + CDEBUG(D_NET, + "Rebinding %s %pI4h from interface %d to %d\n", + libcfs_id2str(peer_ni->ksnp_id), &route->ksnr_ipaddr, - &route->ksnr_myipaddr, - &conn->ksnc_myipaddr); + route->ksnr_myiface, + conn_iface); - iface = ksocknal_ip2iface(route->ksnr_peer->ksnp_ni, - route->ksnr_myipaddr); - if (iface != NULL) - iface->ksni_nroutes--; - } - route->ksnr_myipaddr = conn->ksnc_myipaddr; - iface = ksocknal_ip2iface(route->ksnr_peer->ksnp_ni, - route->ksnr_myipaddr); - if (iface != NULL) - iface->ksni_nroutes++; - } + iface = ksocknal_index2iface(route->ksnr_peer->ksnp_ni, + route->ksnr_myiface); + if (iface) + iface->ksni_nroutes--; + } + route->ksnr_myiface = conn_iface; + iface = ksocknal_index2iface(route->ksnr_peer->ksnp_ni, + route->ksnr_myiface); + if (iface) + iface->ksni_nroutes++; + } - route->ksnr_connected |= (1<ksnr_conn_count++; + route->ksnr_connected |= (1<ksnr_conn_count++; - /* Successful connection => further attempts can - * proceed immediately */ - route->ksnr_retry_interval = 0; + /* Successful connection => further attempts can + * proceed immediately + */ + route->ksnr_retry_interval = 0; } static void @@ -358,12 +411,14 @@ ksocknal_add_route_locked(struct ksock_peer_ni *peer_ni, struct ksock_route *rou struct list_head *tmp; struct ksock_conn *conn; struct ksock_route *route2; + struct ksock_net *net = peer_ni->ksnp_ni->ni_data; LASSERT(!peer_ni->ksnp_closing); LASSERT(route->ksnr_peer == NULL); LASSERT(!route->ksnr_scheduled); LASSERT(!route->ksnr_connecting); LASSERT(route->ksnr_connected == 0); + LASSERT(net->ksnn_ninterfaces > 0); /* LASSERT(unique) */ list_for_each(tmp, &peer_ni->ksnp_routes) { @@ -379,6 +434,11 @@ ksocknal_add_route_locked(struct ksock_peer_ni *peer_ni, struct ksock_route *rou route->ksnr_peer = peer_ni; ksocknal_peer_addref(peer_ni); + + /* set the route's interface to the current net's interface */ + route->ksnr_myiface = net->ksnn_interfaces[0].ksni_index; + net->ksnn_interfaces[0].ksni_nroutes++; + /* peer_ni's routelist takes over my ref on 'route' */ list_add_tail(&route->ksnr_list, &peer_ni->ksnp_routes); @@ -414,10 +474,10 @@ ksocknal_del_route_locked(struct ksock_route *route) ksocknal_close_conn_locked(conn, 0); } - if (route->ksnr_myipaddr != 0) { - iface = ksocknal_ip2iface(route->ksnr_peer->ksnp_ni, - route->ksnr_myipaddr); - if (iface != NULL) + if (route->ksnr_myiface >= 0) { + iface = ksocknal_index2iface(route->ksnr_peer->ksnp_ni, + route->ksnr_myiface); + if (iface) iface->ksni_nroutes--; } @@ -470,8 +530,7 @@ ksocknal_add_peer(struct lnet_ni *ni, struct lnet_process_id id, __u32 ipaddr, peer_ni = peer2; } else { /* peer_ni table takes my ref on peer_ni */ - list_add_tail(&peer_ni->ksnp_list, - ksocknal_nid2peerlist(id.nid)); + hash_add(ksocknal_data.ksnd_peers, &peer_ni->ksnp_list, id.nid); } route2 = NULL; @@ -555,8 +614,7 @@ static int ksocknal_del_peer(struct lnet_ni *ni, struct lnet_process_id id, __u32 ip) { LIST_HEAD(zombies); - struct list_head *ptmp; - struct list_head *pnxt; + struct hlist_node *pnxt; struct ksock_peer_ni *peer_ni; int lo; int hi; @@ -566,19 +624,17 @@ ksocknal_del_peer(struct lnet_ni *ni, struct lnet_process_id id, __u32 ip) write_lock_bh(&ksocknal_data.ksnd_global_lock); if (id.nid != LNET_NID_ANY) { - hi = (int)(ksocknal_nid2peerlist(id.nid) - - ksocknal_data.ksnd_peers); - lo = hi; + lo = hash_min(id.nid, HASH_BITS(ksocknal_data.ksnd_peers)); + hi = lo; } else { lo = 0; - hi = ksocknal_data.ksnd_peer_hash_size - 1; + hi = HASH_SIZE(ksocknal_data.ksnd_peers) - 1; } for (i = lo; i <= hi; i++) { - list_for_each_safe(ptmp, pnxt, - &ksocknal_data.ksnd_peers[i]) { - peer_ni = list_entry(ptmp, struct ksock_peer_ni, ksnp_list); - + hlist_for_each_entry_safe(peer_ni, pnxt, + &ksocknal_data.ksnd_peers[i], + ksnp_list) { if (peer_ni->ksnp_ni != ni) continue; @@ -618,33 +674,27 @@ static struct ksock_conn * ksocknal_get_conn_by_idx(struct lnet_ni *ni, int index) { struct ksock_peer_ni *peer_ni; - struct list_head *ptmp; struct ksock_conn *conn; struct list_head *ctmp; int i; read_lock(&ksocknal_data.ksnd_global_lock); - for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++) { - list_for_each(ptmp, &ksocknal_data.ksnd_peers[i]) { - peer_ni = list_entry(ptmp, struct ksock_peer_ni, ksnp_list); + hash_for_each(ksocknal_data.ksnd_peers, i, peer_ni, ksnp_list) { + LASSERT(!peer_ni->ksnp_closing); - LASSERT(!peer_ni->ksnp_closing); + if (peer_ni->ksnp_ni != ni) + continue; - if (peer_ni->ksnp_ni != ni) + list_for_each(ctmp, &peer_ni->ksnp_conns) { + if (index-- > 0) continue; - list_for_each(ctmp, &peer_ni->ksnp_conns) { - if (index-- > 0) - continue; - - conn = list_entry(ctmp, struct ksock_conn, - ksnc_list); - ksocknal_conn_addref(conn); - read_unlock(&ksocknal_data. \ - ksnd_global_lock); - return conn; - } + conn = list_entry(ctmp, struct ksock_conn, + ksnc_list); + ksocknal_conn_addref(conn); + read_unlock(&ksocknal_data.ksnd_global_lock); + return conn; } } @@ -765,9 +815,10 @@ ksocknal_select_ips(struct ksock_peer_ni *peer_ni, __u32 *peerips, int n_peerips LASSERT(net->ksnn_ninterfaces <= LNET_INTERFACES_NUM); /* Only match interfaces for additional connections - * if I have > 1 interface */ - n_ips = (net->ksnn_ninterfaces < 2) ? 0 : - MIN(n_peerips, net->ksnn_ninterfaces); + * if I have > 1 interface + */ + n_ips = (net->ksnn_ninterfaces < 2) ? 0 : + min(n_peerips, net->ksnn_ninterfaces); for (i = 0; peer_ni->ksnp_n_passive_ips < n_ips; i++) { /* ^ yes really... */ @@ -920,7 +971,7 @@ ksocknal_create_routes(struct ksock_peer_ni *peer_ni, int port, route = list_entry(rtmp, struct ksock_route, ksnr_list); - if (route->ksnr_myipaddr == iface->ksni_ipaddr) + if (route->ksnr_myiface == iface->ksni_index) break; route = NULL; @@ -928,34 +979,34 @@ ksocknal_create_routes(struct ksock_peer_ni *peer_ni, int port, if (route != NULL) continue; - this_netmatch = (((iface->ksni_ipaddr ^ - newroute->ksnr_ipaddr) & - iface->ksni_netmask) == 0) ? 1 : 0; + this_netmatch = (((iface->ksni_ipaddr ^ + newroute->ksnr_ipaddr) & + iface->ksni_netmask) == 0) ? 1 : 0; - if (!(best_iface == NULL || - best_netmatch < this_netmatch || - (best_netmatch == this_netmatch && - best_nroutes > iface->ksni_nroutes))) - continue; + if (!(best_iface == NULL || + best_netmatch < this_netmatch || + (best_netmatch == this_netmatch && + best_nroutes > iface->ksni_nroutes))) + continue; - best_iface = iface; - best_netmatch = this_netmatch; - best_nroutes = iface->ksni_nroutes; - } + best_iface = iface; + best_netmatch = this_netmatch; + best_nroutes = iface->ksni_nroutes; + } - if (best_iface == NULL) - continue; + if (best_iface == NULL) + continue; - newroute->ksnr_myipaddr = best_iface->ksni_ipaddr; - best_iface->ksni_nroutes++; + newroute->ksnr_myiface = best_iface->ksni_index; + best_iface->ksni_nroutes++; - ksocknal_add_route_locked(peer_ni, newroute); - newroute = NULL; - } + ksocknal_add_route_locked(peer_ni, newroute); + newroute = NULL; + } write_unlock_bh(global_lock); - if (newroute != NULL) - ksocknal_route_decref(newroute); + if (newroute != NULL) + ksocknal_route_decref(newroute); } int @@ -1132,8 +1183,8 @@ ksocknal_create_conn(struct lnet_ni *ni, struct ksock_route *route, if (peer2 == NULL) { /* NB this puts an "empty" peer_ni in the peer_ni * table (which takes my ref) */ - list_add_tail(&peer_ni->ksnp_list, - ksocknal_nid2peerlist(peerid.nid)); + hash_add(ksocknal_data.ksnd_peers, + &peer_ni->ksnp_list, peerid.nid); } else { ksocknal_peer_decref(peer_ni); peer_ni = peer2; @@ -1267,7 +1318,7 @@ ksocknal_create_conn(struct lnet_ni *ni, struct ksock_route *route, /* Set the deadline for the outgoing HELLO to drain */ conn->ksnc_tx_bufnob = sock->sk->sk_wmem_queued; conn->ksnc_tx_deadline = ktime_get_seconds() + - lnet_get_lnd_timeout(); + ksocknal_timeout(); smp_mb(); /* order with adding to peer_ni's conn list */ list_add(&conn->ksnc_list, &peer_ni->ksnp_conns); @@ -1356,8 +1407,7 @@ failed_2: if (!peer_ni->ksnp_closing && list_empty(&peer_ni->ksnp_conns) && list_empty(&peer_ni->ksnp_routes)) { - list_add(&zombies, &peer_ni->ksnp_tx_queue); - list_del_init(&peer_ni->ksnp_tx_queue); + list_splice_init(&peer_ni->ksnp_tx_queue, &zombies); ksocknal_unlink_peer_locked(peer_ni); } @@ -1429,7 +1479,7 @@ ksocknal_close_conn_locked(struct ksock_conn *conn, int error) if (route != NULL) { /* dissociate conn from route... */ LASSERT(!route->ksnr_deleted); - LASSERT((route->ksnr_connected & (1 << conn->ksnc_type)) != 0); + LASSERT((route->ksnr_connected & BIT(conn->ksnc_type)) != 0); conn2 = NULL; list_for_each(tmp, &peer_ni->ksnp_conns) { @@ -1442,7 +1492,7 @@ ksocknal_close_conn_locked(struct ksock_conn *conn, int error) conn2 = NULL; } if (conn2 == NULL) - route->ksnr_connected &= ~(1 << conn->ksnc_type); + route->ksnr_connected &= ~BIT(conn->ksnc_type); conn->ksnc_route = NULL; @@ -1538,8 +1588,7 @@ ksocknal_finalize_zcreq(struct ksock_conn *conn) tx->tx_msg.ksm_zc_cookies[0] = 0; tx->tx_zc_aborted = 1; /* mark it as not-acked */ - list_del(&tx->tx_zc_list); - list_add(&tx->tx_zc_list, &zlist); + list_move(&tx->tx_zc_list, &zlist); } spin_unlock(&peer_ni->ksnp_lock); @@ -1646,7 +1695,7 @@ ksocknal_destroy_conn(struct ksock_conn *conn) switch (conn->ksnc_rx_state) { case SOCKNAL_RX_LNET_PAYLOAD: last_rcv = conn->ksnc_rx_deadline - - lnet_get_lnd_timeout(); + ksocknal_timeout(); CERROR("Completing partial receive from %s[%d], " "ip %pI4h:%d, with error, wanted: %d, left: %d, " "last alive is %lld secs ago\n", @@ -1733,8 +1782,7 @@ int ksocknal_close_matching_conns(struct lnet_process_id id, __u32 ipaddr) { struct ksock_peer_ni *peer_ni; - struct list_head *ptmp; - struct list_head *pnxt; + struct hlist_node *pnxt; int lo; int hi; int i; @@ -1742,33 +1790,37 @@ ksocknal_close_matching_conns(struct lnet_process_id id, __u32 ipaddr) write_lock_bh(&ksocknal_data.ksnd_global_lock); - if (id.nid != LNET_NID_ANY) - lo = hi = (int)(ksocknal_nid2peerlist(id.nid) - ksocknal_data.ksnd_peers); - else { - lo = 0; - hi = ksocknal_data.ksnd_peer_hash_size - 1; - } - - for (i = lo; i <= hi; i++) { - list_for_each_safe(ptmp, pnxt, &ksocknal_data.ksnd_peers[i]) { + if (id.nid != LNET_NID_ANY) { + lo = hash_min(id.nid, HASH_BITS(ksocknal_data.ksnd_peers)); + hi = lo; + } else { + lo = 0; + hi = HASH_SIZE(ksocknal_data.ksnd_peers) - 1; + } - peer_ni = list_entry(ptmp, struct ksock_peer_ni, ksnp_list); + for (i = lo; i <= hi; i++) { + hlist_for_each_entry_safe(peer_ni, pnxt, + &ksocknal_data.ksnd_peers[i], + ksnp_list) { - if (!((id.nid == LNET_NID_ANY || id.nid == peer_ni->ksnp_id.nid) && - (id.pid == LNET_PID_ANY || id.pid == peer_ni->ksnp_id.pid))) - continue; + if (!((id.nid == LNET_NID_ANY || + id.nid == peer_ni->ksnp_id.nid) && + (id.pid == LNET_PID_ANY || + id.pid == peer_ni->ksnp_id.pid))) + continue; - count += ksocknal_close_peer_conns_locked (peer_ni, ipaddr, 0); - } - } + count += ksocknal_close_peer_conns_locked(peer_ni, + ipaddr, 0); + } + } write_unlock_bh(&ksocknal_data.ksnd_global_lock); - /* wildcards always succeed */ - if (id.nid == LNET_NID_ANY || id.pid == LNET_PID_ANY || ipaddr == 0) - return (0); + /* wildcards always succeed */ + if (id.nid == LNET_NID_ANY || id.pid == LNET_PID_ANY || ipaddr == 0) + return 0; - return (count == 0 ? -ENOENT : 0); + return (count == 0 ? -ENOENT : 0); } void @@ -1792,69 +1844,6 @@ ksocknal_notify_gw_down(lnet_nid_t gw_nid) * if we have autroutes, and these connect on demand. */ } -void -ksocknal_query(struct lnet_ni *ni, lnet_nid_t nid, time64_t *when) -{ - int connect = 1; - time64_t last_alive = 0; - time64_t now = ktime_get_seconds(); - struct ksock_peer_ni *peer_ni = NULL; - rwlock_t *glock = &ksocknal_data.ksnd_global_lock; - struct lnet_process_id id = { - .nid = nid, - .pid = LNET_PID_LUSTRE, - }; - - read_lock(glock); - - peer_ni = ksocknal_find_peer_locked(ni, id); - if (peer_ni != NULL) { - struct list_head *tmp; - struct ksock_conn *conn; - int bufnob; - - list_for_each(tmp, &peer_ni->ksnp_conns) { - conn = list_entry(tmp, struct ksock_conn, ksnc_list); - bufnob = conn->ksnc_sock->sk->sk_wmem_queued; - - if (bufnob < conn->ksnc_tx_bufnob) { - /* something got ACKed */ - conn->ksnc_tx_deadline = ktime_get_seconds() + - lnet_get_lnd_timeout(); - peer_ni->ksnp_last_alive = now; - conn->ksnc_tx_bufnob = bufnob; - } - } - - last_alive = peer_ni->ksnp_last_alive; - if (ksocknal_find_connectable_route_locked(peer_ni) == NULL) - connect = 0; - } - - read_unlock(glock); - - if (last_alive != 0) - *when = last_alive; - - CDEBUG(D_NET, "peer_ni %s %p, alive %lld secs ago, connect %d\n", - libcfs_nid2str(nid), peer_ni, - last_alive ? now - last_alive : -1, - connect); - - if (!connect) - return; - - ksocknal_add_peer(ni, id, LNET_NIDADDR(nid), lnet_acceptor_port()); - - write_lock_bh(glock); - - peer_ni = ksocknal_find_peer_locked(ni, id); - if (peer_ni != NULL) - ksocknal_launch_all_connections_locked(peer_ni); - - write_unlock_bh(glock); -} - static void ksocknal_push_peer(struct ksock_peer_ni *peer_ni) { @@ -1891,28 +1880,30 @@ ksocknal_push_peer(struct ksock_peer_ni *peer_ni) static int ksocknal_push(struct lnet_ni *ni, struct lnet_process_id id) { - struct list_head *start; - struct list_head *end; - struct list_head *tmp; - int rc = -ENOENT; - unsigned int hsize = ksocknal_data.ksnd_peer_hash_size; + int lo; + int hi; + int bkt; + int rc = -ENOENT; - if (id.nid == LNET_NID_ANY) { - start = &ksocknal_data.ksnd_peers[0]; - end = &ksocknal_data.ksnd_peers[hsize - 1]; + if (id.nid != LNET_NID_ANY) { + lo = hash_min(id.nid, HASH_BITS(ksocknal_data.ksnd_peers)); + hi = lo; } else { - start = end = ksocknal_nid2peerlist(id.nid); + lo = 0; + hi = HASH_SIZE(ksocknal_data.ksnd_peers) - 1; } - for (tmp = start; tmp <= end; tmp++) { - int peer_off; /* searching offset in peer_ni hash table */ + for (bkt = lo; bkt <= hi; bkt++) { + int peer_off; /* searching offset in peer_ni hash table */ for (peer_off = 0; ; peer_off++) { struct ksock_peer_ni *peer_ni; int i = 0; read_lock(&ksocknal_data.ksnd_global_lock); - list_for_each_entry(peer_ni, tmp, ksnp_list) { + hlist_for_each_entry(peer_ni, + &ksocknal_data.ksnd_peers[bkt], + ksnp_list) { if (!((id.nid == LNET_NID_ANY || id.nid == peer_ni->ksnp_id.nid) && (id.pid == LNET_PID_ANY || @@ -1945,7 +1936,6 @@ ksocknal_add_interface(struct lnet_ni *ni, __u32 ipaddress, __u32 netmask) int rc; int i; int j; - struct list_head *ptmp; struct ksock_peer_ni *peer_ni; struct list_head *rtmp; struct ksock_route *route; @@ -1965,33 +1955,32 @@ ksocknal_add_interface(struct lnet_ni *ni, __u32 ipaddress, __u32 netmask) } else { iface = &net->ksnn_interfaces[net->ksnn_ninterfaces++]; + iface->ksni_index = ksocknal_ip2index(ipaddress, ni); iface->ksni_ipaddr = ipaddress; iface->ksni_netmask = netmask; iface->ksni_nroutes = 0; iface->ksni_npeers = 0; - for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++) { - list_for_each(ptmp, &ksocknal_data.ksnd_peers[i]) { - peer_ni = list_entry(ptmp, struct ksock_peer_ni, - ksnp_list); + hash_for_each(ksocknal_data.ksnd_peers, i, peer_ni, ksnp_list) { + for (j = 0; j < peer_ni->ksnp_n_passive_ips; j++) + if (peer_ni->ksnp_passive_ips[j] == ipaddress) + iface->ksni_npeers++; - for (j = 0; j < peer_ni->ksnp_n_passive_ips; j++) - if (peer_ni->ksnp_passive_ips[j] == ipaddress) - iface->ksni_npeers++; - - list_for_each(rtmp, &peer_ni->ksnp_routes) { - route = list_entry(rtmp, - struct ksock_route, - ksnr_list); + list_for_each(rtmp, &peer_ni->ksnp_routes) { + route = list_entry(rtmp, + struct ksock_route, + ksnr_list); - if (route->ksnr_myipaddr == ipaddress) - iface->ksni_nroutes++; - } + if (route->ksnr_myiface == + iface->ksni_index) + iface->ksni_nroutes++; } } rc = 0; - /* NB only new connections will pay attention to the new interface! */ + /* NB only new connections will pay attention to the new + * interface! + */ } write_unlock_bh(&ksocknal_data.ksnd_global_lock); @@ -2000,7 +1989,8 @@ ksocknal_add_interface(struct lnet_ni *ni, __u32 ipaddress, __u32 netmask) } static void -ksocknal_peer_del_interface_locked(struct ksock_peer_ni *peer_ni, __u32 ipaddr) +ksocknal_peer_del_interface_locked(struct ksock_peer_ni *peer_ni, + __u32 ipaddr, int index) { struct list_head *tmp; struct list_head *nxt; @@ -2021,16 +2011,16 @@ ksocknal_peer_del_interface_locked(struct ksock_peer_ni *peer_ni, __u32 ipaddr) list_for_each_safe(tmp, nxt, &peer_ni->ksnp_routes) { route = list_entry(tmp, struct ksock_route, ksnr_list); - if (route->ksnr_myipaddr != ipaddr) - continue; + if (route->ksnr_myiface != index) + continue; - if (route->ksnr_share_count != 0) { - /* Manually created; keep, but unbind */ - route->ksnr_myipaddr = 0; - } else { - ksocknal_del_route_locked(route); - } - } + if (route->ksnr_share_count != 0) { + /* Manually created; keep, but unbind */ + route->ksnr_myiface = -1; + } else { + ksocknal_del_route_locked(route); + } + } list_for_each_safe(tmp, nxt, &peer_ni->ksnp_conns) { conn = list_entry(tmp, struct ksock_conn, ksnc_list); @@ -2045,47 +2035,45 @@ ksocknal_del_interface(struct lnet_ni *ni, __u32 ipaddress) { struct ksock_net *net = ni->ni_data; int rc = -ENOENT; - struct list_head *tmp; - struct list_head *nxt; + struct hlist_node *nxt; struct ksock_peer_ni *peer_ni; u32 this_ip; + int index; int i; int j; - write_lock_bh(&ksocknal_data.ksnd_global_lock); + index = ksocknal_ip2index(ipaddress, ni); - for (i = 0; i < net->ksnn_ninterfaces; i++) { - this_ip = net->ksnn_interfaces[i].ksni_ipaddr; + write_lock_bh(&ksocknal_data.ksnd_global_lock); - if (!(ipaddress == 0 || - ipaddress == this_ip)) - continue; + for (i = 0; i < net->ksnn_ninterfaces; i++) { + this_ip = net->ksnn_interfaces[i].ksni_ipaddr; - rc = 0; + if (!(ipaddress == 0 || + ipaddress == this_ip)) + continue; - for (j = i+1; j < net->ksnn_ninterfaces; j++) - net->ksnn_interfaces[j-1] = - net->ksnn_interfaces[j]; + rc = 0; - net->ksnn_ninterfaces--; + for (j = i+1; j < net->ksnn_ninterfaces; j++) + net->ksnn_interfaces[j-1] = + net->ksnn_interfaces[j]; - for (j = 0; j < ksocknal_data.ksnd_peer_hash_size; j++) { - list_for_each_safe(tmp, nxt, - &ksocknal_data.ksnd_peers[j]) { - peer_ni = list_entry(tmp, struct ksock_peer_ni, - ksnp_list); + net->ksnn_ninterfaces--; - if (peer_ni->ksnp_ni != ni) - continue; + hash_for_each_safe(ksocknal_data.ksnd_peers, j, + nxt, peer_ni, ksnp_list) { + if (peer_ni->ksnp_ni != ni) + continue; - ksocknal_peer_del_interface_locked(peer_ni, this_ip); - } - } - } + ksocknal_peer_del_interface_locked(peer_ni, + this_ip, index); + } + } write_unlock_bh(&ksocknal_data.ksnd_global_lock); - return (rc); + return rc; } int @@ -2223,18 +2211,13 @@ ksocknal_free_buffers (void) if (ksocknal_data.ksnd_schedulers != NULL) cfs_percpt_free(ksocknal_data.ksnd_schedulers); - LIBCFS_FREE (ksocknal_data.ksnd_peers, - sizeof(struct list_head) * - ksocknal_data.ksnd_peer_hash_size); - spin_lock(&ksocknal_data.ksnd_tx_lock); if (!list_empty(&ksocknal_data.ksnd_idle_noop_txs)) { - struct list_head zlist; + LIST_HEAD(zlist); struct ksock_tx *tx; - list_add(&zlist, &ksocknal_data.ksnd_idle_noop_txs); - list_del_init(&ksocknal_data.ksnd_idle_noop_txs); + list_splice_init(&ksocknal_data.ksnd_idle_noop_txs, &zlist); spin_unlock(&ksocknal_data.ksnd_tx_lock); while (!list_empty(&zlist)) { @@ -2251,6 +2234,7 @@ static void ksocknal_base_shutdown(void) { struct ksock_sched *sched; + struct ksock_peer_ni *peer_ni; int i; CDEBUG(D_MALLOC, "before NAL cleanup: kmem %d\n", @@ -2264,9 +2248,8 @@ ksocknal_base_shutdown(void) case SOCKNAL_INIT_ALL: case SOCKNAL_INIT_DATA: - LASSERT(ksocknal_data.ksnd_peers != NULL); - for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++) - LASSERT(list_empty(&ksocknal_data.ksnd_peers[i])); + hash_for_each(ksocknal_data.ksnd_peers, i, peer_ni, ksnp_list) + LASSERT(0); LASSERT(list_empty(&ksocknal_data.ksnd_nets)); LASSERT(list_empty(&ksocknal_data.ksnd_enomem_conns)); @@ -2296,26 +2279,16 @@ ksocknal_base_shutdown(void) wake_up_all(&sched->kss_waitq); } - i = 4; - read_lock(&ksocknal_data.ksnd_global_lock); - while (ksocknal_data.ksnd_nthreads != 0) { - i++; - /* power of 2? */ - CDEBUG(((i & (-i)) == i) ? D_WARNING : D_NET, - "waiting for %d threads to terminate\n", - ksocknal_data.ksnd_nthreads); - read_unlock(&ksocknal_data.ksnd_global_lock); - set_current_state(TASK_UNINTERRUPTIBLE); - schedule_timeout(cfs_time_seconds(1)); - read_lock(&ksocknal_data.ksnd_global_lock); - } - read_unlock(&ksocknal_data.ksnd_global_lock); + wait_var_event_warning(&ksocknal_data.ksnd_nthreads, + ksocknal_data.ksnd_nthreads == 0, + "waiting for %d threads to terminate\n", + ksocknal_data.ksnd_nthreads); - ksocknal_free_buffers(); + ksocknal_free_buffers(); - ksocknal_data.ksnd_init = SOCKNAL_INIT_NOTHING; - break; - } + ksocknal_data.ksnd_init = SOCKNAL_INIT_NOTHING; + break; + } CDEBUG(D_MALLOC, "after NAL cleanup: kmem %d\n", atomic_read (&libcfs_kmemory)); @@ -2330,20 +2303,12 @@ ksocknal_base_startup(void) int rc; int i; - LASSERT (ksocknal_data.ksnd_init == SOCKNAL_INIT_NOTHING); - LASSERT (ksocknal_data.ksnd_nnets == 0); - - memset (&ksocknal_data, 0, sizeof (ksocknal_data)); /* zero pointers */ + LASSERT(ksocknal_data.ksnd_init == SOCKNAL_INIT_NOTHING); + LASSERT(ksocknal_data.ksnd_nnets == 0); - ksocknal_data.ksnd_peer_hash_size = SOCKNAL_PEER_HASH_SIZE; - LIBCFS_ALLOC(ksocknal_data.ksnd_peers, - sizeof(struct list_head) * - ksocknal_data.ksnd_peer_hash_size); - if (ksocknal_data.ksnd_peers == NULL) - return -ENOMEM; + memset(&ksocknal_data, 0, sizeof(ksocknal_data)); /* zero pointers */ - for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++) - INIT_LIST_HEAD(&ksocknal_data.ksnd_peers[i]); + hash_init(ksocknal_data.ksnd_peers); rwlock_init(&ksocknal_data.ksnd_global_lock); INIT_LIST_HEAD(&ksocknal_data.ksnd_nets); @@ -2366,7 +2331,8 @@ ksocknal_base_startup(void) /* flag lists/ptrs/locks initialised */ ksocknal_data.ksnd_init = SOCKNAL_INIT_DATA; - try_module_get(THIS_MODULE); + if (!try_module_get(THIS_MODULE)) + goto failed; /* Create a scheduler block per available CPT */ ksocknal_data.ksnd_schedulers = cfs_percpt_alloc(lnet_cpt_table(), @@ -2451,29 +2417,20 @@ ksocknal_base_startup(void) return -ENETDOWN; } -static void +static int ksocknal_debug_peerhash(struct lnet_ni *ni) { - struct ksock_peer_ni *peer_ni = NULL; - struct list_head *tmp; + struct ksock_peer_ni *peer_ni; int i; read_lock(&ksocknal_data.ksnd_global_lock); - for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++) { - list_for_each(tmp, &ksocknal_data.ksnd_peers[i]) { - peer_ni = list_entry(tmp, struct ksock_peer_ni, ksnp_list); - - if (peer_ni->ksnp_ni == ni) - break; - - peer_ni = NULL; - } - } - - if (peer_ni != NULL) { + hash_for_each(ksocknal_data.ksnd_peers, i, peer_ni, ksnp_list) { struct ksock_route *route; - struct ksock_conn *conn; + struct ksock_conn *conn; + + if (peer_ni->ksnp_ni != ni) + continue; CWARN("Active peer_ni on shutdown: %s, ref %d, " "closing %d, accepting %d, err %d, zcookie %llu, " @@ -2485,24 +2442,24 @@ ksocknal_debug_peerhash(struct lnet_ni *ni) !list_empty(&peer_ni->ksnp_tx_queue), !list_empty(&peer_ni->ksnp_zc_req_list)); - list_for_each(tmp, &peer_ni->ksnp_routes) { - route = list_entry(tmp, struct ksock_route, ksnr_list); + list_for_each_entry(route, &peer_ni->ksnp_routes, ksnr_list) { CWARN("Route: ref %d, schd %d, conn %d, cnted %d, " "del %d\n", atomic_read(&route->ksnr_refcount), route->ksnr_scheduled, route->ksnr_connecting, route->ksnr_connected, route->ksnr_deleted); } - list_for_each(tmp, &peer_ni->ksnp_conns) { - conn = list_entry(tmp, struct ksock_conn, ksnc_list); + list_for_each_entry(conn, &peer_ni->ksnp_conns, ksnc_list) { CWARN("Conn: ref %d, sref %d, t %d, c %d\n", atomic_read(&conn->ksnc_conn_refcount), atomic_read(&conn->ksnc_sock_refcount), conn->ksnc_type, conn->ksnc_closing); } + break; } read_unlock(&ksocknal_data.ksnd_global_lock); + return 0; } void @@ -2525,17 +2482,13 @@ ksocknal_shutdown(struct lnet_ni *ni) ksocknal_del_peer(ni, anyid, 0); /* Wait for all peer_ni state to clean up */ - i = 2; - while (atomic_read(&net->ksnn_npeers) > SOCKNAL_SHUTDOWN_BIAS) { - i++; - CDEBUG(((i & (-i)) == i) ? D_WARNING : D_NET, /* power of 2? */ - "waiting for %d peers to disconnect\n", - atomic_read(&net->ksnn_npeers) - SOCKNAL_SHUTDOWN_BIAS); - set_current_state(TASK_UNINTERRUPTIBLE); - schedule_timeout(cfs_time_seconds(1)); - - ksocknal_debug_peerhash(ni); - } + wait_var_event_warning(&net->ksnn_npeers, + atomic_read(&net->ksnn_npeers) == + SOCKNAL_SHUTDOWN_BIAS, + "waiting for %d peers to disconnect\n", + ksocknal_debug_peerhash(ni) + + atomic_read(&net->ksnn_npeers) - + SOCKNAL_SHUTDOWN_BIAS); for (i = 0; i < net->ksnn_ninterfaces; i++) { LASSERT(net->ksnn_interfaces[i].ksni_npeers == 0); @@ -2545,9 +2498,9 @@ ksocknal_shutdown(struct lnet_ni *ni) list_del(&net->ksnn_list); LIBCFS_FREE(net, sizeof(*net)); - ksocknal_data.ksnd_nnets--; - if (ksocknal_data.ksnd_nnets == 0) - ksocknal_base_shutdown(); + ksocknal_data.ksnd_nnets--; + if (ksocknal_data.ksnd_nnets == 0) + ksocknal_base_shutdown(); } static int @@ -2722,6 +2675,7 @@ ksocknal_startup(struct lnet_ni *ni) net->ksnn_ninterfaces = 1; ni->ni_dev_cpt = ifaces[0].li_cpt; ksi->ksni_ipaddr = ifaces[0].li_ipaddr; + ksi->ksni_index = ksocknal_ip2index(ksi->ksni_ipaddr, ni); ksi->ksni_netmask = ifaces[0].li_netmask; strlcpy(ksi->ksni_name, ifaces[0].li_name, sizeof(ksi->ksni_name)); @@ -2761,6 +2715,8 @@ ksocknal_startup(struct lnet_ni *ni) ksi = &net->ksnn_interfaces[j]; ni->ni_dev_cpt = ifaces[j].li_cpt; ksi->ksni_ipaddr = ifaces[j].li_ipaddr; + ksi->ksni_index = + ksocknal_ip2index(ksi->ksni_ipaddr, ni); ksi->ksni_netmask = ifaces[j].li_netmask; strlcpy(ksi->ksni_name, ifaces[j].li_name, sizeof(ksi->ksni_name)); @@ -2804,6 +2760,17 @@ static void __exit ksocklnd_exit(void) lnet_unregister_lnd(&the_ksocklnd); } +static const struct lnet_lnd the_ksocklnd = { + .lnd_type = SOCKLND, + .lnd_startup = ksocknal_startup, + .lnd_shutdown = ksocknal_shutdown, + .lnd_ctl = ksocknal_ctl, + .lnd_send = ksocknal_send, + .lnd_recv = ksocknal_recv, + .lnd_notify_peer_down = ksocknal_notify_gw_down, + .lnd_accept = ksocknal_accept, +}; + static int __init ksocklnd_init(void) { int rc; @@ -2812,17 +2779,6 @@ static int __init ksocklnd_init(void) BUILD_BUG_ON(SOCKLND_CONN_NTYPES > 4); BUILD_BUG_ON(SOCKLND_CONN_ACK != SOCKLND_CONN_BULK_IN); - /* initialize the_ksocklnd */ - the_ksocklnd.lnd_type = SOCKLND; - the_ksocklnd.lnd_startup = ksocknal_startup; - the_ksocklnd.lnd_shutdown = ksocknal_shutdown; - the_ksocklnd.lnd_ctl = ksocknal_ctl; - the_ksocklnd.lnd_send = ksocknal_send; - the_ksocklnd.lnd_recv = ksocknal_recv; - the_ksocklnd.lnd_notify_peer_down = ksocknal_notify_gw_down; - the_ksocklnd.lnd_query = ksocknal_query; - the_ksocklnd.lnd_accept = ksocknal_accept; - rc = ksocknal_tunables_init(); if (rc != 0) return rc;