Whamcloud - gitweb
LU-1346 gnilnd: remove libcfs abstractions
[fs/lustre-release.git] / lnet / klnds / o2iblnd / o2iblnd_cb.c
index 76c844d..5c9d1a1 100644 (file)
@@ -26,6 +26,8 @@
 /*
  * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
  * Use is subject to license terms.
+ *
+ * Copyright (c) 2012, Intel Corporation.
  */
 /*
  * This file is part of Lustre, http://www.lustre.org/
@@ -549,7 +551,7 @@ kiblnd_kvaddr_to_page (unsigned long vaddr)
 static int
 kiblnd_fmr_map_tx(kib_net_t *net, kib_tx_t *tx, kib_rdma_desc_t *rd, int nob)
 {
-       kib_hca_dev_t           *hdev  = tx->tx_pool->tpo_hdev;
+       kib_hca_dev_t           *hdev;
        __u64                   *pages = tx->tx_pages;
        kib_fmr_poolset_t       *fps;
        int                     npages;
@@ -558,6 +560,11 @@ kiblnd_fmr_map_tx(kib_net_t *net, kib_tx_t *tx, kib_rdma_desc_t *rd, int nob)
        int                     rc;
        int                     i;
 
+       LASSERT(tx->tx_pool != NULL);
+       LASSERT(tx->tx_pool->tpo_pool.po_owner != NULL);
+
+       hdev  = tx->tx_pool->tpo_hdev;
+
         for (i = 0, npages = 0; i < rd->rd_nfrags; i++) {
                 for (size = 0; size <  rd->rd_frags[i].rf_nob;
                                size += hdev->ibh_page_size) {
@@ -566,9 +573,6 @@ kiblnd_fmr_map_tx(kib_net_t *net, kib_tx_t *tx, kib_rdma_desc_t *rd, int nob)
                 }
         }
 
-       LASSERT(tx->tx_pool != NULL);
-       LASSERT(tx->tx_pool->tpo_pool.po_owner != NULL);
-
        cpt = tx->tx_pool->tpo_pool.po_owner->ps_cpt;
 
        fps = net->ibn_fmr_ps[cpt];
@@ -592,17 +596,19 @@ kiblnd_fmr_map_tx(kib_net_t *net, kib_tx_t *tx, kib_rdma_desc_t *rd, int nob)
 static int
 kiblnd_pmr_map_tx(kib_net_t *net, kib_tx_t *tx, kib_rdma_desc_t *rd, int nob)
 {
-       kib_hca_dev_t           *hdev = tx->tx_pool->tpo_hdev;
+       kib_hca_dev_t           *hdev;
        kib_pmr_poolset_t       *pps;
        __u64                   iova;
        int                     cpt;
        int                     rc;
 
-       iova = rd->rd_frags[0].rf_addr & ~hdev->ibh_page_mask;
-
        LASSERT(tx->tx_pool != NULL);
        LASSERT(tx->tx_pool->tpo_pool.po_owner != NULL);
 
+       hdev = tx->tx_pool->tpo_hdev;
+
+       iova = rd->rd_frags[0].rf_addr & ~hdev->ibh_page_mask;
+
        cpt = tx->tx_pool->tpo_pool.po_owner->ps_cpt;
 
        pps = net->ibn_pmr_ps[cpt];
@@ -1170,8 +1176,8 @@ kiblnd_queue_tx_locked (kib_tx_t *tx, kib_conn_t *conn)
         LASSERT (!tx->tx_queued);               /* not queued for sending already */
         LASSERT (conn->ibc_state >= IBLND_CONN_ESTABLISHED);
 
-        tx->tx_queued = 1;
-        tx->tx_deadline = jiffies + (*kiblnd_tunables.kib_timeout * CFS_HZ);
+       tx->tx_queued = 1;
+       tx->tx_deadline = jiffies + (*kiblnd_tunables.kib_timeout * HZ);
 
         if (tx->tx_conn == NULL) {
                 kiblnd_conn_addref(conn);
@@ -1799,15 +1805,15 @@ kiblnd_recv (lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg, int delayed,
 }
 
 int
-kiblnd_thread_start (int (*fn)(void *arg), void *arg)
+kiblnd_thread_start(int (*fn)(void *arg), void *arg, char *name)
 {
-        long    pid = cfs_create_thread (fn, arg, 0);
+       struct task_struct *task = kthread_run(fn, arg, name);
 
-        if (pid < 0)
-                return ((int)pid);
+       if (IS_ERR(task))
+               return PTR_ERR(task);
 
-        cfs_atomic_inc (&kiblnd_data.kib_nthreads);
-        return (0);
+       cfs_atomic_inc(&kiblnd_data.kib_nthreads);
+       return 0;
 }
 
 void
@@ -1903,17 +1909,17 @@ kiblnd_close_conn_locked (kib_conn_t *conn, int error)
 
         kiblnd_set_conn_state(conn, IBLND_CONN_CLOSING);
 
-        if (error != 0 &&
-            kiblnd_dev_can_failover(dev)) {
-                cfs_list_add_tail(&dev->ibd_fail_list,
-                              &kiblnd_data.kib_failed_devs);
-                cfs_waitq_signal(&kiblnd_data.kib_failover_waitq);
-        }
+       if (error != 0 &&
+           kiblnd_dev_can_failover(dev)) {
+               cfs_list_add_tail(&dev->ibd_fail_list,
+                             &kiblnd_data.kib_failed_devs);
+               wake_up(&kiblnd_data.kib_failover_waitq);
+       }
 
        spin_lock_irqsave(&kiblnd_data.kib_connd_lock, flags);
 
        cfs_list_add_tail(&conn->ibc_list, &kiblnd_data.kib_connd_conns);
-       cfs_waitq_signal(&kiblnd_data.kib_connd_waitq);
+       wake_up(&kiblnd_data.kib_connd_waitq);
 
        spin_unlock_irqrestore(&kiblnd_data.kib_connd_lock, flags);
 }
@@ -3120,20 +3126,19 @@ kiblnd_disconnect_conn (kib_conn_t *conn)
 int
 kiblnd_connd (void *arg)
 {
-        cfs_waitlink_t     wait;
-        unsigned long      flags;
-        kib_conn_t        *conn;
-        int                timeout;
-        int                i;
-        int                dropped_lock;
-        int                peer_index = 0;
-        unsigned long      deadline = jiffies;
+       wait_queue_t     wait;
+       unsigned long      flags;
+       kib_conn_t        *conn;
+       int                timeout;
+       int                i;
+       int                dropped_lock;
+       int                peer_index = 0;
+       unsigned long      deadline = jiffies;
 
-        cfs_daemonize ("kiblnd_connd");
-        cfs_block_allsigs ();
+       cfs_block_allsigs ();
 
-        cfs_waitlink_init (&wait);
-        kiblnd_data.kib_connd = current;
+       init_waitqueue_entry_current (&wait);
+       kiblnd_data.kib_connd = current;
 
        spin_lock_irqsave(&kiblnd_data.kib_connd_lock, flags);
 
@@ -3195,13 +3200,13 @@ kiblnd_connd (void *arg)
                         if (chunk == 0)
                                 chunk = 1;
 
-                        for (i = 0; i < chunk; i++) {
-                                kiblnd_check_conns(peer_index);
-                                peer_index = (peer_index + 1) %
-                                             kiblnd_data.kib_peer_hash_size;
-                        }
+                       for (i = 0; i < chunk; i++) {
+                               kiblnd_check_conns(peer_index);
+                               peer_index = (peer_index + 1) %
+                                            kiblnd_data.kib_peer_hash_size;
+                       }
 
-                        deadline += p * CFS_HZ;
+                       deadline += p * HZ;
                        spin_lock_irqsave(&kiblnd_data.kib_connd_lock, flags);
                }
 
@@ -3209,14 +3214,14 @@ kiblnd_connd (void *arg)
                        continue;
 
                /* Nothing to do for 'timeout'  */
-               cfs_set_current_state(CFS_TASK_INTERRUPTIBLE);
-               cfs_waitq_add(&kiblnd_data.kib_connd_waitq, &wait);
+               set_current_state(TASK_INTERRUPTIBLE);
+               add_wait_queue(&kiblnd_data.kib_connd_waitq, &wait);
                spin_unlock_irqrestore(&kiblnd_data.kib_connd_lock, flags);
 
-               cfs_waitq_timedwait(&wait, CFS_TASK_INTERRUPTIBLE, timeout);
+               waitq_timedwait(&wait, TASK_INTERRUPTIBLE, timeout);
 
-               cfs_set_current_state(CFS_TASK_RUNNING);
-               cfs_waitq_del(&kiblnd_data.kib_connd_waitq, &wait);
+               set_current_state(TASK_RUNNING);
+               remove_wait_queue(&kiblnd_data.kib_connd_waitq, &wait);
                spin_lock_irqsave(&kiblnd_data.kib_connd_lock, flags);
        }
 
@@ -3298,8 +3303,8 @@ kiblnd_cq_completion(struct ib_cq *cq, void *arg)
                conn->ibc_scheduled = 1;
                cfs_list_add_tail(&conn->ibc_sched_list, &sched->ibs_conns);
 
-               if (cfs_waitq_active(&sched->ibs_waitq))
-                       cfs_waitq_signal(&sched->ibs_waitq);
+               if (waitqueue_active(&sched->ibs_waitq))
+                       wake_up(&sched->ibs_waitq);
        }
 
        spin_unlock_irqrestore(&sched->ibs_lock, flags);
@@ -3320,30 +3325,25 @@ kiblnd_scheduler(void *arg)
        long                    id = (long)arg;
        struct kib_sched_info   *sched;
        kib_conn_t              *conn;
-       cfs_waitlink_t          wait;
+       wait_queue_t            wait;
        unsigned long           flags;
        struct ib_wc            wc;
-       char                    name[20];
        int                     did_something;
        int                     busy_loops = 0;
        int                     rc;
 
-       snprintf(name, sizeof(name), "kiblnd_sd_%02ld_%02ld",
-                KIB_THREAD_CPT(id), KIB_THREAD_TID(id));
-
-       cfs_daemonize(name);
        cfs_block_allsigs();
 
-       cfs_waitlink_init(&wait);
+       init_waitqueue_entry_current(&wait);
 
        sched = kiblnd_data.kib_scheds[KIB_THREAD_CPT(id)];
 
        rc = cfs_cpt_bind(lnet_cpt_table(), sched->ibs_cpt);
        if (rc != 0) {
-               CWARN("Failed to bind %s on CPT %d, please verify whether "
+               CWARN("Failed to bind on CPT %d, please verify whether "
                      "all CPUs are healthy and reload modules if necessary, "
                      "otherwise your system might under risk of low "
-                     "performance\n", name, sched->ibs_cpt);
+                     "performance\n", sched->ibs_cpt);
        }
 
        spin_lock_irqsave(&sched->ibs_lock, flags);
@@ -3352,7 +3352,7 @@ kiblnd_scheduler(void *arg)
                if (busy_loops++ >= IBLND_RESCHED) {
                        spin_unlock_irqrestore(&sched->ibs_lock, flags);
 
-                       cfs_cond_resched();
+                       cond_resched();
                        busy_loops = 0;
 
                        spin_lock_irqsave(&sched->ibs_lock, flags);
@@ -3409,8 +3409,8 @@ kiblnd_scheduler(void *arg)
                                kiblnd_conn_addref(conn);
                                cfs_list_add_tail(&conn->ibc_sched_list,
                                                  &sched->ibs_conns);
-                               if (cfs_waitq_active(&sched->ibs_waitq))
-                                       cfs_waitq_signal(&sched->ibs_waitq);
+                               if (waitqueue_active(&sched->ibs_waitq))
+                                       wake_up(&sched->ibs_waitq);
                        } else {
                                conn->ibc_scheduled = 0;
                        }
@@ -3429,15 +3429,15 @@ kiblnd_scheduler(void *arg)
                 if (did_something)
                         continue;
 
-               cfs_set_current_state(CFS_TASK_INTERRUPTIBLE);
-               cfs_waitq_add_exclusive(&sched->ibs_waitq, &wait);
+               set_current_state(TASK_INTERRUPTIBLE);
+               add_wait_queue_exclusive(&sched->ibs_waitq, &wait);
                spin_unlock_irqrestore(&sched->ibs_lock, flags);
 
-               cfs_waitq_wait(&wait, CFS_TASK_INTERRUPTIBLE);
+               waitq_wait(&wait, TASK_INTERRUPTIBLE);
                busy_loops = 0;
 
-               cfs_waitq_del(&sched->ibs_waitq, &wait);
-               cfs_set_current_state(CFS_TASK_RUNNING);
+               remove_wait_queue(&sched->ibs_waitq, &wait);
+               set_current_state(TASK_RUNNING);
                spin_lock_irqsave(&sched->ibs_lock, flags);
        }
 
@@ -3451,17 +3451,16 @@ int
 kiblnd_failover_thread(void *arg)
 {
        rwlock_t                *glock = &kiblnd_data.kib_global_lock;
-        kib_dev_t         *dev;
-        cfs_waitlink_t     wait;
-        unsigned long      flags;
-        int                rc;
+       kib_dev_t         *dev;
+       wait_queue_t     wait;
+       unsigned long      flags;
+       int                rc;
 
-        LASSERT (*kiblnd_tunables.kib_dev_failover != 0);
+       LASSERT (*kiblnd_tunables.kib_dev_failover != 0);
 
-        cfs_daemonize ("kiblnd_failover");
-        cfs_block_allsigs ();
+       cfs_block_allsigs ();
 
-        cfs_waitlink_init(&wait);
+       init_waitqueue_entry_current(&wait);
        write_lock_irqsave(glock, flags);
 
         while (!kiblnd_data.kib_shutdown) {
@@ -3507,14 +3506,14 @@ kiblnd_failover_thread(void *arg)
                 /* long sleep if no more pending failover */
                 long_sleep = cfs_list_empty(&kiblnd_data.kib_failed_devs);
 
-                cfs_set_current_state(CFS_TASK_INTERRUPTIBLE);
-                cfs_waitq_add(&kiblnd_data.kib_failover_waitq, &wait);
+               set_current_state(TASK_INTERRUPTIBLE);
+               add_wait_queue(&kiblnd_data.kib_failover_waitq, &wait);
                write_unlock_irqrestore(glock, flags);
 
                rc = schedule_timeout(long_sleep ? cfs_time_seconds(10) :
                                                   cfs_time_seconds(1));
-               cfs_set_current_state(CFS_TASK_RUNNING);
-               cfs_waitq_del(&kiblnd_data.kib_failover_waitq, &wait);
+               set_current_state(TASK_RUNNING);
+               remove_wait_queue(&kiblnd_data.kib_failover_waitq, &wait);
                write_lock_irqsave(glock, flags);
 
                 if (!long_sleep || rc != 0)