*/
#include "ranal.h"
+static int kranal_devids[] = {RAPK_MAIN_DEVICE_ID,
+ RAPK_EXPANSION_DEVICE_ID};
nal_t kranal_api;
ptl_handle_ni_t kranal_ni;
#define RANAL_SYSCTL 202
static ctl_table kranal_ctl_table[] = {
- {RANAL_SYSCTL_TIMEOUT, "timeout",
+ {RANAL_SYSCTL_TIMEOUT, "timeout",
&kranal_tunables.kra_timeout, sizeof(int),
0644, NULL, &proc_dointvec},
- {RANAL_SYSCTL_LISTENER_TIMEOUT, "listener_timeout",
+ {RANAL_SYSCTL_LISTENER_TIMEOUT, "listener_timeout",
&kranal_tunables.kra_listener_timeout, sizeof(int),
0644, NULL, &proc_dointvec},
{RANAL_SYSCTL_BACKLOG, "backlog",
{RANAL_SYSCTL_PORT, "port",
&kranal_tunables.kra_port, sizeof(int),
0644, NULL, kranal_listener_procint},
- {RANAL_SYSCTL_MAX_IMMEDIATE, "max_immediate",
+ {RANAL_SYSCTL_MAX_IMMEDIATE, "max_immediate",
&kranal_tunables.kra_max_immediate, sizeof(int),
0644, NULL, &proc_dointvec},
{ 0 }
if (rc == nob)
return 0;
-
+
if (rc >= 0)
return -EAGAIN;
int rc;
mm_segment_t oldmm = get_fs();
long ticks = timeout * HZ;
- int wanted = nob;
unsigned long then;
struct timeval tv;
ticks -= jiffies - then;
set_fs(oldmm);
- CDEBUG(D_WARNING, "rc %d at %d/%d bytes %d/%d secs\n",
- rc, wanted - nob, wanted, timeout - (int)(ticks/HZ), timeout);
-
if (rc < 0)
return rc;
rrc = RapkGetRiParams(conn->rac_rihandle, &connreq->racr_riparams);
LASSERT(rrc == RAP_SUCCESS);
-
- CDEBUG(D_WARNING,"devid %d, riparams: HID %08x FDH %08x PT %08x CC %08x\n",
- connreq->racr_devid,
- connreq->racr_riparams.HostId,
- connreq->racr_riparams.FmaDomainHndl,
- connreq->racr_riparams.PTag,
- connreq->racr_riparams.CompletionCookie);
}
int
connreq->racr_timeout, RANAL_MIN_TIMEOUT);
return -EPROTO;
}
-
+
return 0;
}
if (conn->rac_device != newconn->rac_device)
continue;
-
+
if (loopback &&
newconn->rac_my_connstamp == conn->rac_peer_connstamp &&
newconn->rac_peer_connstamp == conn->rac_my_connstamp)
continue;
-
+
LASSERT (conn->rac_peer_connstamp < newconn->rac_peer_connstamp);
CDEBUG(D_NET, "Closing stale conn nid:"LPX64
- " connstamp:"LPX64"("LPX64")\n", peer->rap_nid,
+ " connstamp:"LPX64"("LPX64")\n", peer->rap_nid,
conn->rac_peer_connstamp, newconn->rac_peer_connstamp);
count++;
int loopback;
loopback = peer->rap_nid == kranal_lib.libnal_ni.ni_pid.nid;
-
+
list_for_each(tmp, &peer->rap_conns) {
conn = list_entry(tmp, kra_conn_t, rac_list);
/* 'newconn' is an earlier connection from 'peer'!!! */
if (newconn->rac_peer_connstamp < conn->rac_peer_connstamp)
return 2;
-
+
/* 'conn' is an earlier connection from 'peer': it will be
* removed when we cull stale conns later on... */
if (newconn->rac_peer_connstamp > conn->rac_peer_connstamp)
continue;
-
+
/* 'newconn' has the SAME connection stamp; 'peer' isn't
* playing the game... */
return 3;
do { /* allocate a unique cqid */
conn->rac_cqid = kranal_data.kra_next_cqid++;
} while (kranal_cqid2conn_locked(conn->rac_cqid) != NULL);
-
+
write_unlock_irqrestore(&kranal_data.kra_global_lock, flags);
}
}
void
-kranal_destroy_conn(kra_conn_t *conn)
+kranal_destroy_conn(kra_conn_t *conn)
{
RAP_RETURN rrc;
/* Non-persistent peer with no more conns... */
kranal_unlink_peer_locked(peer);
}
-
+
/* Reset RX timeout to ensure we wait for an incoming CLOSE for the
- * full timeout */
+ * full timeout. If we get a CLOSE we know the peer has stopped all
+ * RDMA. Otherwise if we wait for the full timeout we can also be sure
+ * all RDMA has stopped. */
conn->rac_last_rx = jiffies;
mb();
kranal_close_conn (kra_conn_t *conn, int error)
{
unsigned long flags;
-
+
write_lock_irqsave(&kranal_data.kra_global_lock, flags);
-
+
if (conn->rac_state == RANAL_CONN_ESTABLISHED)
kranal_close_conn_locked(conn, error);
-
+
write_unlock_irqrestore(&kranal_data.kra_global_lock, flags);
}
int
-kranal_set_conn_params(kra_conn_t *conn, kra_connreq_t *connreq,
+kranal_set_conn_params(kra_conn_t *conn, kra_connreq_t *connreq,
__u32 peer_ip, int peer_port)
{
- RAP_RETURN rrc;
-
- CDEBUG(D_WARNING,"devid %d, riparams: HID %08x FDH %08x PT %08x CC %08x\n",
- conn->rac_device->rad_id,
- connreq->racr_riparams.HostId,
- connreq->racr_riparams.FmaDomainHndl,
- connreq->racr_riparams.PTag,
- connreq->racr_riparams.CompletionCookie);
-
+ kra_device_t *dev = conn->rac_device;
+ unsigned long flags;
+ RAP_RETURN rrc;
+
+ /* CAVEAT EMPTOR: we're really overloading rac_last_tx + rac_keepalive
+ * to do RapkCompleteSync() timekeeping (see kibnal_scheduler). */
+ conn->rac_last_tx = jiffies;
+ conn->rac_keepalive = 0;
+
+ /* Schedule conn on rad_new_conns */
+ kranal_conn_addref(conn);
+ spin_lock_irqsave(&dev->rad_lock, flags);
+ list_add_tail(&conn->rac_schedlist, &dev->rad_new_conns);
+ wake_up(&dev->rad_waitq);
+ spin_unlock_irqrestore(&dev->rad_lock, flags);
+
rrc = RapkSetRiParams(conn->rac_rihandle, &connreq->racr_riparams);
if (rrc != RAP_SUCCESS) {
- CERROR("Error setting riparams from %u.%u.%u.%u/%d: %d\n",
+ CERROR("Error setting riparams from %u.%u.%u.%u/%d: %d\n",
HIPQUAD(peer_ip), peer_port, rrc);
- return -EPROTO;
+ return -ECONNABORTED;
}
-
+
+ /* Scheduler doesn't touch conn apart from to deschedule and decref it
+ * after RapkCompleteSync() return success, so conn is all mine */
+
conn->rac_peerstamp = connreq->racr_peerstamp;
conn->rac_peer_connstamp = connreq->racr_connstamp;
conn->rac_keepalive = RANAL_TIMEOUT2KEEPALIVE(connreq->racr_timeout);
}
int
-kranal_passive_conn_handshake (struct socket *sock, ptl_nid_t *src_nidp,
+kranal_passive_conn_handshake (struct socket *sock, ptl_nid_t *src_nidp,
ptl_nid_t *dst_nidp, kra_conn_t **connp)
{
struct sockaddr_in addr;
int len;
int i;
- CDEBUG(D_WARNING,"!!\n");
-
len = sizeof(addr);
rc = sock->ops->getname(sock, (struct sockaddr *)&addr, &len, 2);
if (rc != 0) {
peer_ip = ntohl(addr.sin_addr.s_addr);
peer_port = ntohs(addr.sin_port);
- CDEBUG(D_WARNING,"%u.%u.%u.%u\n", HIPQUAD(peer_ip));
-
if (peer_port >= 1024) {
CERROR("Refusing unprivileged connection from %u.%u.%u.%u/%d\n",
HIPQUAD(peer_ip), peer_port);
return -ECONNREFUSED;
}
- CDEBUG(D_WARNING,"%u.%u.%u.%u\n", HIPQUAD(peer_ip));
-
- rc = kranal_recv_connreq(sock, &rx_connreq,
+ rc = kranal_recv_connreq(sock, &rx_connreq,
kranal_tunables.kra_listener_timeout);
if (rc != 0) {
- CERROR("Can't rx connreq from %u.%u.%u.%u/%d: %d\n",
+ CERROR("Can't rx connreq from %u.%u.%u.%u/%d: %d\n",
HIPQUAD(peer_ip), peer_port, rc);
return rc;
}
- CDEBUG(D_WARNING,"%u.%u.%u.%u\n", HIPQUAD(peer_ip));
-
for (i = 0;;i++) {
if (i == kranal_data.kra_ndevs) {
CERROR("Can't match dev %d from %u.%u.%u.%u/%d\n",
break;
}
- CDEBUG(D_WARNING,"%u.%u.%u.%u\n", HIPQUAD(peer_ip));
-
rc = kranal_create_conn(&conn, dev);
if (rc != 0)
return rc;
- CDEBUG(D_WARNING,"%u.%u.%u.%u\n", HIPQUAD(peer_ip));
-
kranal_pack_connreq(&tx_connreq, conn, rx_connreq.racr_srcnid);
rc = kranal_sock_write(sock, &tx_connreq, sizeof(tx_connreq));
if (rc != 0) {
- CERROR("Can't tx connreq to %u.%u.%u.%u/%d: %d\n",
+ CERROR("Can't tx connreq to %u.%u.%u.%u/%d: %d\n",
HIPQUAD(peer_ip), peer_port, rc);
kranal_conn_decref(conn);
return rc;
}
- CDEBUG(D_WARNING,"%u.%u.%u.%u\n", HIPQUAD(peer_ip));
-
rc = kranal_set_conn_params(conn, &rx_connreq, peer_ip, peer_port);
if (rc != 0) {
kranal_conn_decref(conn);
return rc;
}
- CDEBUG(D_WARNING,"%u.%u.%u.%u\n", HIPQUAD(peer_ip));
-
*connp = conn;
*src_nidp = rx_connreq.racr_srcnid;
*dst_nidp = rx_connreq.racr_dstnid;
for (port = 1023; port >= 512; port--) {
- memset(&locaddr, 0, sizeof(locaddr));
- locaddr.sin_family = AF_INET;
+ memset(&locaddr, 0, sizeof(locaddr));
+ locaddr.sin_family = AF_INET;
locaddr.sin_port = htons(port);
locaddr.sin_addr.s_addr = htonl(INADDR_ANY);
(struct sockaddr *)&locaddr, sizeof(locaddr));
if (rc != 0) {
sock_release(sock);
-
+
if (rc == -EADDRINUSE) {
CDEBUG(D_NET, "Port %d already in use\n", port);
continue;
*sockp = sock;
return 0;
}
-
+
sock_release(sock);
if (rc != -EADDRNOTAVAIL) {
port, HIPQUAD(peer->rap_ip), peer->rap_port, rc);
return rc;
}
-
- CDEBUG(D_NET, "Port %d not available for %u.%u.%u.%u/%d\n",
+
+ CDEBUG(D_NET, "Port %d not available for %u.%u.%u.%u/%d\n",
port, HIPQUAD(peer->rap_ip), peer->rap_port);
}
int
-kranal_active_conn_handshake(kra_peer_t *peer,
+kranal_active_conn_handshake(kra_peer_t *peer,
ptl_nid_t *dst_nidp, kra_conn_t **connp)
{
kra_connreq_t connreq;
int rc;
unsigned int idx;
- CDEBUG(D_WARNING,LPX64"\n", peer->rap_nid);
-
/* spread connections over all devices using both peer NIDs to ensure
* all nids use all devices */
idx = peer->rap_nid + kranal_lib.libnal_ni.ni_pid.nid;
if (rc != 0)
return rc;
- CDEBUG(D_WARNING,LPX64"\n", peer->rap_nid);
-
kranal_pack_connreq(&connreq, conn, peer->rap_nid);
-
+
rc = ranal_connect_sock(peer, &sock);
if (rc != 0)
goto failed_0;
- CDEBUG(D_WARNING,LPX64"\n", peer->rap_nid);
-
/* CAVEAT EMPTOR: the passive side receives with a SHORT rx timeout
* immediately after accepting a connection, so we connect and then
* send immediately. */
rc = kranal_sock_write(sock, &connreq, sizeof(connreq));
if (rc != 0) {
- CERROR("Can't tx connreq to %u.%u.%u.%u/%d: %d\n",
+ CERROR("Can't tx connreq to %u.%u.%u.%u/%d: %d\n",
HIPQUAD(peer->rap_ip), peer->rap_port, rc);
goto failed_1;
}
- CDEBUG(D_WARNING,LPX64"\n", peer->rap_nid);
-
rc = kranal_recv_connreq(sock, &connreq, kranal_tunables.kra_timeout);
if (rc != 0) {
- CERROR("Can't rx connreq from %u.%u.%u.%u/%d: %d\n",
+ CERROR("Can't rx connreq from %u.%u.%u.%u/%d: %d\n",
HIPQUAD(peer->rap_ip), peer->rap_port, rc);
goto failed_1;
}
- CDEBUG(D_WARNING,LPX64"\n", peer->rap_nid);
-
sock_release(sock);
rc = -EPROTO;
if (connreq.racr_srcnid != peer->rap_nid) {
CERROR("Unexpected srcnid from %u.%u.%u.%u/%d: "
"received "LPX64" expected "LPX64"\n",
- HIPQUAD(peer->rap_ip), peer->rap_port,
+ HIPQUAD(peer->rap_ip), peer->rap_port,
connreq.racr_srcnid, peer->rap_nid);
goto failed_0;
}
if (connreq.racr_devid != dev->rad_id) {
CERROR("Unexpected device id from %u.%u.%u.%u/%d: "
"received %d expected %d\n",
- HIPQUAD(peer->rap_ip), peer->rap_port,
+ HIPQUAD(peer->rap_ip), peer->rap_port,
connreq.racr_devid, dev->rad_id);
goto failed_0;
}
- CDEBUG(D_WARNING,LPX64"\n", peer->rap_nid);
-
- rc = kranal_set_conn_params(conn, &connreq,
+ rc = kranal_set_conn_params(conn, &connreq,
peer->rap_ip, peer->rap_port);
if (rc != 0)
goto failed_0;
*connp = conn;
*dst_nidp = connreq.racr_dstnid;
- CDEBUG(D_WARNING,LPX64"\n", peer->rap_nid);
return 0;
failed_1:
sock_release(sock);
failed_0:
kranal_conn_decref(conn);
- CDEBUG(D_WARNING,LPX64": %d\n", peer->rap_nid, rc);
return rc;
}
/* active: connd wants to connect to 'peer' */
LASSERT (peer != NULL);
LASSERT (peer->rap_connecting);
-
+
rc = kranal_active_conn_handshake(peer, &dst_nid, &conn);
if (rc != 0)
return rc;
if (!kranal_peer_active(peer)) {
/* raced with peer getting unlinked */
- write_unlock_irqrestore(&kranal_data.kra_global_lock,
+ write_unlock_irqrestore(&kranal_data.kra_global_lock,
flags);
kranal_conn_decref(conn);
return -ESTALE;
}
}
- LASSERT (!new_peer == !kranal_peer_active(peer));
+ LASSERT ((!new_peer) != (!kranal_peer_active(peer)));
/* Refuse connection if peer thinks we are a different NID. We check
* this while holding the global lock, to synch with connection
list_add_tail(&peer->rap_list,
kranal_nid2peerlist(peer_nid));
}
-
+
+ /* initialise timestamps before reaper looks at them */
+ conn->rac_last_tx = conn->rac_last_rx = jiffies;
+
kranal_peer_addref(peer); /* +1 ref for conn */
conn->rac_peer = peer;
list_add_tail(&conn->rac_list, &peer->rap_conns);
if (nstale != 0)
CWARN("Closed %d stale conns to "LPX64"\n", nstale, peer_nid);
+ CDEBUG(D_WARNING, "New connection to "LPX64" on devid[%d] = %d\n",
+ peer_nid, conn->rac_device->rad_idx, conn->rac_device->rad_id);
+
/* Ensure conn gets checked. Transmits may have been queued and an
* FMA event may have happened before it got in the cq hash table */
kranal_schedule_conn(conn);
LASSERT (peer->rap_connecting);
- CDEBUG(D_WARNING,"About to handshake "LPX64"\n", peer->rap_nid);
+ CDEBUG(D_NET, "About to handshake "LPX64"\n", peer->rap_nid);
rc = kranal_conn_handshake(NULL, peer);
- CDEBUG(D_WARNING,"Done handshake "LPX64":%d \n", peer->rap_nid, rc);
+ CDEBUG(D_NET, "Done handshake "LPX64":%d \n", peer->rap_nid, rc);
write_lock_irqsave(&kranal_data.kra_global_lock, flags);
rc = sock->ops->listen(sock, kranal_tunables.kra_backlog);
if (rc != 0) {
- CERROR("Can't set listen backlog %d: %d\n",
+ CERROR("Can't set listen backlog %d: %d\n",
kranal_tunables.kra_backlog, rc);
goto out_1;
}
ras->ras_sock->type = sock->type;
ras->ras_sock->ops = sock->ops;
}
-
+
set_current_state(TASK_INTERRUPTIBLE);
rc = sock->ops->accept(sock, ras->ras_sock, O_NONBLOCK);
if (rc == 0) {
spin_lock_irqsave(&kranal_data.kra_connd_lock, flags);
-
- list_add_tail(&ras->ras_list,
+
+ list_add_tail(&ras->ras_list,
&kranal_data.kra_connd_acceptq);
spin_unlock_irqrestore(&kranal_data.kra_connd_lock, flags);
ras = NULL;
continue;
}
-
+
if (rc != -EAGAIN) {
CERROR("Accept failed: %d, pausing...\n", rc);
kranal_pause(HZ);
sock_release(sock);
kranal_data.kra_listener_sock = NULL;
out_0:
- /* set completion status and unblock thread waiting for me
+ /* set completion status and unblock thread waiting for me
* (parent on startup failure, executioner on normal shutdown) */
kranal_data.kra_listener_shutdown = rc;
up(&kranal_data.kra_listener_signal);
long pid;
int rc;
- CDEBUG(D_WARNING, "Starting listener\n");
+ CDEBUG(D_NET, "Starting listener\n");
/* Called holding kra_nid_mutex: listener stopped */
LASSERT (kranal_data.kra_listener_sock == NULL);
rc = kranal_data.kra_listener_shutdown;
LASSERT ((rc != 0) == (kranal_data.kra_listener_sock == NULL));
- CDEBUG(D_WARNING, "Listener %ld started OK\n", pid);
+ CDEBUG(D_NET, "Listener %ld started OK\n", pid);
return rc;
}
unsigned long flags;
kra_acceptsock_t *ras;
- CDEBUG(D_WARNING, "Stopping listener\n");
+ CDEBUG(D_NET, "Stopping listener\n");
/* Called holding kra_nid_mutex: listener running */
LASSERT (kranal_data.kra_listener_sock != NULL);
down(&kranal_data.kra_listener_signal);
LASSERT (kranal_data.kra_listener_sock == NULL);
- CDEBUG(D_WARNING, "Listener stopped\n");
+ CDEBUG(D_NET, "Listener stopped\n");
if (!clear_acceptq)
return;
-
+
/* Close any unhandled accepts */
spin_lock_irqsave(&kranal_data.kra_connd_lock, flags);
list_del_init(&kranal_data.kra_connd_acceptq);
spin_unlock_irqrestore(&kranal_data.kra_connd_lock, flags);
-
+
while (!list_empty(&zombie_accepts)) {
- ras = list_entry(zombie_accepts.next,
+ ras = list_entry(zombie_accepts.next,
kra_acceptsock_t, ras_list);
list_del(&ras->ras_list);
kranal_free_acceptsock(ras);
}
}
-int
+int
kranal_listener_procint(ctl_table *table, int write, struct file *filp,
void *buffer, size_t *lenp)
{
kranal_data.kra_peerstamp++;
ni->ni_pid.nid = nid;
write_unlock_irqrestore(&kranal_data.kra_global_lock, flags);
-
+
/* Delete all existing peers and their connections after new
* NID/connstamp set to ensure no old connections in our brave
* new world. */
}
int
-kranal_get_peer_info (int index, ptl_nid_t *nidp, __u32 *ipp, int *portp,
+kranal_get_peer_info (int index, ptl_nid_t *nidp, __u32 *ipp, int *portp,
int *persistencep)
{
kra_peer_t *peer;
break;
}
case NAL_CMD_DEL_PEER: {
- rc = kranal_del_peer(pcfg->pcfg_nid,
+ rc = kranal_del_peer(pcfg->pcfg_nid,
/* flags == single_share */
pcfg->pcfg_flags != 0);
break;
PORTAL_ALLOC(tx->tx_phys,
PTL_MD_MAX_IOV * sizeof(*tx->tx_phys));
if (tx->tx_phys == NULL) {
- CERROR("Can't allocate %stx[%d]->tx_phys\n",
+ CERROR("Can't allocate %stx[%d]->tx_phys\n",
isnblk ? "nblk " : "", i);
PORTAL_FREE(tx, sizeof(*tx));
{
int i;
unsigned long flags;
-
+
if (nal->nal_refct != 0) {
/* This module got the first ref */
PORTAL_MODULE_UNUSE;
break;
}
+ /* Conn/Peer state all cleaned up BEFORE setting shutdown, so threads
+ * don't have to worry about shutdown races */
+ LASSERT (atomic_read(&kranal_data.kra_nconns) == 0);
+ LASSERT (atomic_read(&kranal_data.kra_npeers) == 0);
+
/* flag threads to terminate; wake and wait for them to die */
kranal_data.kra_shutdown = 1;
for (i = 0; i < kranal_data.kra_ndevs; i++) {
kra_device_t *dev = &kranal_data.kra_devices[i];
- LASSERT (list_empty(&dev->rad_connq));
+ LASSERT (list_empty(&dev->rad_ready_conns));
+ LASSERT (list_empty(&dev->rad_new_conns));
spin_lock_irqsave(&dev->rad_lock, flags);
wake_up(&dev->rad_waitq);
spin_unlock_irqrestore(&kranal_data.kra_reaper_lock, flags);
LASSERT (list_empty(&kranal_data.kra_connd_peers));
- spin_lock_irqsave(&kranal_data.kra_connd_lock, flags);
+ spin_lock_irqsave(&kranal_data.kra_connd_lock, flags);
wake_up_all(&kranal_data.kra_connd_waitq);
- spin_unlock_irqrestore(&kranal_data.kra_connd_lock, flags);
+ spin_unlock_irqrestore(&kranal_data.kra_connd_lock, flags);
i = 2;
while (atomic_read(&kranal_data.kra_nthreads) != 0) {
LASSERT (list_empty(&kranal_data.kra_peers[i]));
PORTAL_FREE(kranal_data.kra_peers,
- sizeof (struct list_head) *
+ sizeof (struct list_head) *
kranal_data.kra_peer_hash_size);
}
LASSERT (list_empty(&kranal_data.kra_conns[i]));
PORTAL_FREE(kranal_data.kra_conns,
- sizeof (struct list_head) *
+ sizeof (struct list_head) *
kranal_data.kra_conn_hash_size);
}
ptl_ni_limits_t *requested_limits,
ptl_ni_limits_t *actual_limits)
{
- static int device_ids[] = {RAPK_MAIN_DEVICE_ID,
- RAPK_EXPANSION_DEVICE_ID};
struct timeval tv;
ptl_process_id_t process_id;
int pkmem = atomic_read(&portal_kmemory);
kra_device_t *dev = &kranal_data.kra_devices[i];
dev->rad_idx = i;
- INIT_LIST_HEAD(&dev->rad_connq);
+ INIT_LIST_HEAD(&dev->rad_ready_conns);
+ INIT_LIST_HEAD(&dev->rad_new_conns);
init_waitqueue_head(&dev->rad_waitq);
spin_lock_init(&dev->rad_lock);
}
/* OK to call kranal_api_shutdown() to cleanup now */
kranal_data.kra_init = RANAL_INIT_DATA;
-
+
kranal_data.kra_peer_hash_size = RANAL_PEER_HASH_SIZE;
PORTAL_ALLOC(kranal_data.kra_peers,
sizeof(struct list_head) * kranal_data.kra_peer_hash_size);
}
}
- LASSERT(kranal_data.kra_ndevs == 0);
- for (i = 0; i < sizeof(device_ids)/sizeof(device_ids[0]); i++) {
+ LASSERT (kranal_data.kra_ndevs == 0);
+
+ for (i = 0; i < sizeof(kranal_devids)/sizeof(kranal_devids[0]); i++) {
+ LASSERT (i < RANAL_MAXDEVS);
+
dev = &kranal_data.kra_devices[kranal_data.kra_ndevs];
- rc = kranal_device_init(device_ids[i], dev);
+ rc = kranal_device_init(kranal_devids[i], dev);
if (rc == 0)
kranal_data.kra_ndevs++;
-
+ }
+
+ if (kranal_data.kra_ndevs == 0) {
+ CERROR("Can't initialise any RapidArray devices\n");
+ goto failed;
+ }
+
+ for (i = 0; i < kranal_data.kra_ndevs; i++) {
rc = kranal_thread_start(kranal_scheduler, dev);
if (rc != 0) {
CERROR("Can't spawn ranal scheduler[%d]: %d\n",
}
}
- if (kranal_data.kra_ndevs == 0)
- goto failed;
-
rc = libcfs_nal_cmd_register(RANAL, &kranal_cmd, NULL);
if (rc != 0) {
CERROR("Can't initialise command interface (rc = %d)\n", rc);
return PTL_OK;
failed:
- kranal_api_shutdown(&kranal_api);
+ kranal_api_shutdown(&kranal_api);
return PTL_FAIL;
}
return -ENODEV;
}
- kranal_tunables.kra_sysctl =
+ kranal_tunables.kra_sysctl =
register_sysctl_table(kranal_top_ctl_table, 0);
if (kranal_tunables.kra_sysctl == NULL) {
CERROR("Can't register sysctl table\n");