LIBCFS_FREE (route, sizeof (*route));
}
-static int
-ksocknal_create_peer(struct ksock_peer_ni **peerp, struct lnet_ni *ni,
- struct lnet_process_id id)
+static struct ksock_peer_ni *
+ksocknal_create_peer(struct lnet_ni *ni, struct lnet_process_id id)
{
int cpt = lnet_cpt_of_nid(id.nid, ni);
struct ksock_net *net = ni->ni_data;
LIBCFS_CPT_ALLOC(peer_ni, lnet_cpt_table(), cpt, sizeof(*peer_ni));
if (peer_ni == NULL)
- return -ENOMEM;
+ return ERR_PTR(-ENOMEM);
peer_ni->ksnp_ni = ni;
peer_ni->ksnp_id = id;
LIBCFS_FREE(peer_ni, sizeof(*peer_ni));
CERROR("Can't create peer_ni: network shutdown\n");
- return -ESHUTDOWN;
+ return ERR_PTR(-ESHUTDOWN);
}
net->ksnn_npeers++;
spin_unlock_bh(&net->ksnn_lock);
- *peerp = peer_ni;
- return 0;
+ return peer_ni;
}
void
struct ksock_peer_ni *peer2;
struct ksock_route *route;
struct ksock_route *route2;
- int rc;
if (id.nid == LNET_NID_ANY ||
id.pid == LNET_PID_ANY)
return (-EINVAL);
- /* Have a brand new peer_ni ready... */
- rc = ksocknal_create_peer(&peer_ni, ni, id);
- if (rc != 0)
- return rc;
+ /* Have a brand new peer_ni ready... */
+ peer_ni = ksocknal_create_peer(ni, id);
+ if (IS_ERR(peer_ni))
+ return PTR_ERR(peer_ni);
route = ksocknal_create_route (ipaddr, port);
if (route == NULL) {
static int
ksocknal_del_peer(struct lnet_ni *ni, struct lnet_process_id id, __u32 ip)
{
- struct list_head zombies = LIST_HEAD_INIT(zombies);
+ LIST_HEAD(zombies);
struct list_head *ptmp;
struct list_head *pnxt;
struct ksock_peer_ni *peer_ni;
struct socket *sock, int type)
{
rwlock_t *global_lock = &ksocknal_data.ksnd_global_lock;
- struct list_head zombies = LIST_HEAD_INIT(zombies);
+ LIST_HEAD(zombies);
struct lnet_process_id peerid;
struct list_head *tmp;
u64 incarnation;
cpt = lnet_cpt_of_nid(peerid.nid, ni);
- if (active) {
- ksocknal_peer_addref(peer_ni);
+ if (active) {
+ ksocknal_peer_addref(peer_ni);
write_lock_bh(global_lock);
- } else {
- rc = ksocknal_create_peer(&peer_ni, ni, peerid);
- if (rc != 0)
- goto failed_1;
+ } else {
+ peer_ni = ksocknal_create_peer(ni, peerid);
+ if (IS_ERR(peer_ni)) {
+ rc = PTR_ERR(peer_ni);
+ goto failed_1;
+ }
write_lock_bh(global_lock);
- /* called with a ref on ni, so shutdown can't have started */
+ /* called with a ref on ni, so shutdown can't have started */
LASSERT(((struct ksock_net *) ni->ni_data)->ksnn_shutdown == 0);
peer2 = ksocknal_find_peer_locked(ni, peerid);
/* No more connections to this peer_ni */
if (!list_empty(&peer_ni->ksnp_tx_queue)) {
- struct ksock_tx *tx;
+ struct ksock_tx *tx;
LASSERT(conn->ksnc_proto == &ksocknal_protocol_v3x);
spin_lock_bh(&ksocknal_data.ksnd_reaper_lock);
- list_add_tail(&conn->ksnc_list,
- &ksocknal_data.ksnd_deathrow_conns);
+ list_add_tail(&conn->ksnc_list, &ksocknal_data.ksnd_deathrow_conns);
wake_up(&ksocknal_data.ksnd_reaper_waitq);
spin_unlock_bh(&ksocknal_data.ksnd_reaper_lock);
struct ksock_peer_ni *peer_ni = conn->ksnc_peer;
struct ksock_tx *tx;
struct ksock_tx *tmp;
- struct list_head zlist = LIST_HEAD_INIT(zlist);
+ LIST_HEAD(zlist);
/* NB safe to finalize TXs because closing of socket will
* abort all buffered data */
ksocknal_launch_all_connections_locked(peer_ni);
write_unlock_bh(glock);
- return;
}
static void
}
read_unlock(&ksocknal_data.ksnd_global_lock);
- if (i == 0) /* no match */
+ if (i <= peer_off) /* no match */
break;
rc = 0;
atomic_read (&libcfs_kmemory));
LASSERT (ksocknal_data.ksnd_nnets == 0);
- switch (ksocknal_data.ksnd_init) {
- default:
- LASSERT (0);
+ switch (ksocknal_data.ksnd_init) {
+ default:
+ LASSERT(0);
+ /* fallthrough */
- case SOCKNAL_INIT_ALL:
- case SOCKNAL_INIT_DATA:
- LASSERT (ksocknal_data.ksnd_peers != NULL);
- for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++) {
+ case SOCKNAL_INIT_ALL:
+ case SOCKNAL_INIT_DATA:
+ LASSERT(ksocknal_data.ksnd_peers != NULL);
+ for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++)
LASSERT(list_empty(&ksocknal_data.ksnd_peers[i]));
- }
LASSERT(list_empty(&ksocknal_data.ksnd_nets));
LASSERT(list_empty(&ksocknal_data.ksnd_enomem_conns));
read_lock(&ksocknal_data.ksnd_global_lock);
- for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++) {
+ for (i = 0; i < ksocknal_data.ksnd_peer_hash_size; i++) {
list_for_each(tmp, &ksocknal_data.ksnd_peers[i]) {
peer_ni = list_entry(tmp, struct ksock_peer_ni, ksnp_list);
- if (peer_ni->ksnp_ni == ni) break;
+ if (peer_ni->ksnp_ni == ni)
+ break;
- peer_ni = NULL;
- }
- }
+ peer_ni = NULL;
+ }
+ }
if (peer_ni != NULL) {
struct ksock_route *route;
struct ksock_conn *conn;
- CWARN ("Active peer_ni on shutdown: %s, ref %d, scnt %d, "
- "closing %d, accepting %d, err %d, zcookie %llu, "
- "txq %d, zc_req %d\n", libcfs_id2str(peer_ni->ksnp_id),
- atomic_read(&peer_ni->ksnp_refcount),
- peer_ni->ksnp_sharecount, peer_ni->ksnp_closing,
- peer_ni->ksnp_accepting, peer_ni->ksnp_error,
- peer_ni->ksnp_zc_next_cookie,
- !list_empty(&peer_ni->ksnp_tx_queue),
- !list_empty(&peer_ni->ksnp_zc_req_list));
+ CWARN("Active peer_ni on shutdown: %s, ref %d, "
+ "closing %d, accepting %d, err %d, zcookie %llu, "
+ "txq %d, zc_req %d\n", libcfs_id2str(peer_ni->ksnp_id),
+ atomic_read(&peer_ni->ksnp_refcount),
+ peer_ni->ksnp_closing,
+ peer_ni->ksnp_accepting, peer_ni->ksnp_error,
+ peer_ni->ksnp_zc_next_cookie,
+ !list_empty(&peer_ni->ksnp_tx_queue),
+ !list_empty(&peer_ni->ksnp_zc_req_list));
list_for_each(tmp, &peer_ni->ksnp_routes) {
route = list_entry(tmp, struct ksock_route, ksnr_list);
- CWARN ("Route: ref %d, schd %d, conn %d, cnted %d, "
- "del %d\n", atomic_read(&route->ksnr_refcount),
- route->ksnr_scheduled, route->ksnr_connecting,
- route->ksnr_connected, route->ksnr_deleted);
+ CWARN("Route: ref %d, schd %d, conn %d, cnted %d, "
+ "del %d\n", atomic_read(&route->ksnr_refcount),
+ route->ksnr_scheduled, route->ksnr_connecting,
+ route->ksnr_connected, route->ksnr_deleted);
}
list_for_each(tmp, &peer_ni->ksnp_conns) {
conn = list_entry(tmp, struct ksock_conn, ksnc_list);
- CWARN ("Conn: ref %d, sref %d, t %d, c %d\n",
- atomic_read(&conn->ksnc_conn_refcount),
- atomic_read(&conn->ksnc_sock_refcount),
- conn->ksnc_type, conn->ksnc_closing);
+ CWARN("Conn: ref %d, sref %d, t %d, c %d\n",
+ atomic_read(&conn->ksnc_conn_refcount),
+ atomic_read(&conn->ksnc_sock_refcount),
+ conn->ksnc_type, conn->ksnc_closing);
}
}
read_unlock(&ksocknal_data.ksnd_global_lock);
- return;
}
void
net_tunables->lct_peer_rtr_credits =
*ksocknal_tunables.ksnd_peerrtrcredits;
- rc = lnet_inet_enumerate(&ifaces);
+ rc = lnet_inet_enumerate(&ifaces, ni->ni_net_ns);
if (rc < 0)
goto fail_1;
int rc;
/* check ksnr_connected/connecting field large enough */
- CLASSERT(SOCKLND_CONN_NTYPES <= 4);
- CLASSERT(SOCKLND_CONN_ACK == SOCKLND_CONN_BULK_IN);
+ BUILD_BUG_ON(SOCKLND_CONN_NTYPES > 4);
+ BUILD_BUG_ON(SOCKLND_CONN_ACK != SOCKLND_CONN_BULK_IN);
/* initialize the_ksocklnd */
the_ksocklnd.lnd_type = SOCKLND;