Whamcloud - gitweb
LU-9679 lnet: use LIST_HEAD() for local lists.
[fs/lustre-release.git] / lnet / lnet / net_fault.c
index 1810121..3d78ce4 100644 (file)
@@ -21,7 +21,7 @@
  * GPL HEADER END
  */
 /*
- * Copyright (c) 2014, Intel Corporation.
+ * Copyright (c) 2014, 2017, Intel Corporation.
  */
 /*
  * This file is part of Lustre, http://www.lustre.org/
@@ -36,8 +36,9 @@
 
 #define DEBUG_SUBSYSTEM S_LNET
 
+#include <linux/random.h>
 #include <lnet/lib-lnet.h>
-#include <lnet/lnetctl.h>
+#include <uapi/linux/lnet/lnetctl.h>
 
 #define LNET_MSG_MASK          (LNET_PUT_BIT | LNET_ACK_BIT | \
                                 LNET_GET_BIT | LNET_REPLY_BIT)
@@ -57,9 +58,9 @@ struct lnet_drop_rule {
        /**
         * seconds to drop the next message, it's exclusive with dr_drop_at
         */
-       cfs_time_t              dr_drop_time;
+       time64_t                dr_drop_time;
        /** baseline to caculate dr_drop_time */
-       cfs_time_t              dr_time_base;
+       time64_t                dr_time_base;
        /** statistic of dropped messages */
        struct lnet_fault_stat  dr_stat;
 };
@@ -79,10 +80,12 @@ lnet_fault_nid_match(lnet_nid_t nid, lnet_nid_t msg_nid)
 
 static bool
 lnet_fault_attr_match(struct lnet_fault_attr *attr, lnet_nid_t src,
-                     lnet_nid_t dst, unsigned int type, unsigned int portal)
+                     lnet_nid_t local_nid, lnet_nid_t dst,
+                     unsigned int type, unsigned int portal)
 {
        if (!lnet_fault_nid_match(attr->fa_src, src) ||
-           !lnet_fault_nid_match(attr->fa_dst, dst))
+           !lnet_fault_nid_match(attr->fa_dst, dst) ||
+           !lnet_fault_nid_match(attr->fa_local_nid, local_nid))
                return false;
 
        if (!(attr->fa_msg_mask & (1 << type)))
@@ -137,6 +140,10 @@ lnet_fault_stat_inc(struct lnet_fault_stat *stat, unsigned int type)
 }
 
 /**
+ * LNet message drop simulation
+ */
+
+/**
  * Add a new drop rule to LNet
  * There is no check for duplicated drop rule, all rules will be checked for
  * incoming message.
@@ -147,8 +154,10 @@ lnet_drop_rule_add(struct lnet_fault_attr *attr)
        struct lnet_drop_rule *rule;
        ENTRY;
 
-       if ((attr->u.drop.da_rate == 0) == (attr->u.drop.da_interval == 0)) {
-               CDEBUG(D_NET, "invalid rate %d or interval %d\n",
+       if (!((attr->u.drop.da_rate == 0) ^ (attr->u.drop.da_interval == 0))) {
+               CDEBUG(D_NET,
+                      "please provide either drop rate or drop interval, "
+                      "but not both at the same time %d/%d\n",
                       attr->u.drop.da_rate, attr->u.drop.da_interval);
                RETURN(-EINVAL);
        }
@@ -164,11 +173,11 @@ lnet_drop_rule_add(struct lnet_fault_attr *attr)
 
        rule->dr_attr = *attr;
        if (attr->u.drop.da_interval != 0) {
-               rule->dr_time_base = cfs_time_shift(attr->u.drop.da_interval);
-               rule->dr_drop_time = cfs_time_shift(cfs_rand() %
-                                                   attr->u.drop.da_interval);
+               rule->dr_time_base = ktime_get_seconds() + attr->u.drop.da_interval;
+               rule->dr_drop_time = ktime_get_seconds() +
+                                    prandom_u32_max(attr->u.drop.da_interval);
        } else {
-               rule->dr_drop_at = cfs_rand() % attr->u.drop.da_rate;
+               rule->dr_drop_at = prandom_u32_max(attr->u.drop.da_rate);
        }
 
        lnet_net_lock(LNET_LOCK_EX);
@@ -193,12 +202,10 @@ lnet_drop_rule_del(lnet_nid_t src, lnet_nid_t dst)
 {
        struct lnet_drop_rule *rule;
        struct lnet_drop_rule *tmp;
-       struct list_head       zombies;
-       int                    n = 0;
+       LIST_HEAD(zombies);
+       int n = 0;
        ENTRY;
 
-       INIT_LIST_HEAD(&zombies);
-
        lnet_net_lock(LNET_LOCK_EX);
        list_for_each_entry_safe(rule, tmp, &the_lnet.ln_drop_rules, dr_link) {
                if (rule->dr_attr.fa_src != src && src != 0)
@@ -275,12 +282,11 @@ lnet_drop_rule_reset(void)
 
                memset(&rule->dr_stat, 0, sizeof(rule->dr_stat));
                if (attr->u.drop.da_rate != 0) {
-                       rule->dr_drop_at = cfs_rand() % attr->u.drop.da_rate;
+                       rule->dr_drop_at = prandom_u32_max(attr->u.drop.da_rate);
                } else {
-                       rule->dr_drop_time = cfs_time_shift(cfs_rand() %
-                                               attr->u.drop.da_interval);
-                       rule->dr_time_base = cfs_time_shift(attr->u.drop.
-                                                                 da_interval);
+                       rule->dr_drop_time = ktime_get_seconds() +
+                                            prandom_u32_max(attr->u.drop.da_interval);
+                       rule->dr_time_base = ktime_get_seconds() + attr->u.drop.da_interval;
                }
                spin_unlock(&rule->dr_lock);
        }
@@ -289,57 +295,125 @@ lnet_drop_rule_reset(void)
        EXIT;
 }
 
+static void
+lnet_fault_match_health(enum lnet_msg_hstatus *hstatus, __u32 mask)
+{
+       int choice;
+       int delta;
+       int best_delta;
+       int i;
+
+       /* assign a random failure */
+       choice = prandom_u32_max(LNET_MSG_STATUS_END - LNET_MSG_STATUS_OK);
+       if (choice == 0)
+               choice++;
+
+       if (mask == HSTATUS_RANDOM) {
+               *hstatus = choice;
+               return;
+       }
+
+       if (mask & (1 << choice)) {
+               *hstatus = choice;
+               return;
+       }
+
+       /* round to the closest ON bit */
+       i = HSTATUS_END;
+       best_delta = HSTATUS_END;
+       while (i > 0) {
+               if (mask & (1 << i)) {
+                       delta = choice - i;
+                       if (delta < 0)
+                               delta *= -1;
+                       if (delta < best_delta) {
+                               best_delta = delta;
+                               choice = i;
+                       }
+               }
+               i--;
+       }
+
+       *hstatus = choice;
+}
+
 /**
  * check source/destination NID, portal, message type and drop rate,
  * decide whether should drop this message or not
  */
 static bool
 drop_rule_match(struct lnet_drop_rule *rule, lnet_nid_t src,
-               lnet_nid_t dst, unsigned int type, unsigned int portal)
+               lnet_nid_t local_nid, lnet_nid_t dst,
+               unsigned int type, unsigned int portal,
+               enum lnet_msg_hstatus *hstatus)
 {
        struct lnet_fault_attr  *attr = &rule->dr_attr;
        bool                     drop;
 
-       if (!lnet_fault_attr_match(attr, src, dst, type, portal))
+       if (!lnet_fault_attr_match(attr, src, local_nid, dst, type, portal))
+               return false;
+
+       if (attr->u.drop.da_drop_all) {
+               CDEBUG(D_NET, "set to drop all messages\n");
+               drop = true;
+               goto drop_matched;
+       }
+
+       /*
+        * if we're trying to match a health status error but it hasn't
+        * been set in the rule, then don't match
+        */
+       if ((hstatus && !attr->u.drop.da_health_error_mask) ||
+           (!hstatus && attr->u.drop.da_health_error_mask))
                return false;
 
        /* match this rule, check drop rate now */
        spin_lock(&rule->dr_lock);
-       if (rule->dr_drop_time != 0) { /* time based drop */
-               cfs_time_t now = cfs_time_current();
+       if (attr->u.drop.da_random) {
+               int value = prandom_u32_max(attr->u.drop.da_interval);
+               if (value >= (attr->u.drop.da_interval / 2))
+                       drop = true;
+               else
+                       drop = false;
+       } else if (rule->dr_drop_time != 0) { /* time based drop */
+               time64_t now = ktime_get_seconds();
 
                rule->dr_stat.fs_count++;
-               drop = cfs_time_aftereq(now, rule->dr_drop_time);
+               drop = now >= rule->dr_drop_time;
                if (drop) {
-                       if (cfs_time_after(now, rule->dr_time_base))
+                       if (now > rule->dr_time_base)
                                rule->dr_time_base = now;
 
                        rule->dr_drop_time = rule->dr_time_base +
-                                            cfs_time_seconds(cfs_rand() %
-                                               attr->u.drop.da_interval);
-                       rule->dr_time_base += cfs_time_seconds(attr->u.drop.
-                                                              da_interval);
-
-                       CDEBUG(D_NET, "Drop Rule %s->%s: next drop : "
-                                     CFS_TIME_T"\n",
-                                     libcfs_nid2str(attr->fa_src),
-                                     libcfs_nid2str(attr->fa_dst),
-                                     rule->dr_drop_time);
+                                            prandom_u32_max(attr->u.drop.da_interval);
+                       rule->dr_time_base += attr->u.drop.da_interval;
+
+                       CDEBUG(D_NET, "Drop Rule %s->%s: next drop : %lld\n",
+                              libcfs_nid2str(attr->fa_src),
+                              libcfs_nid2str(attr->fa_dst),
+                              rule->dr_drop_time);
                }
 
        } else { /* rate based drop */
-               drop = rule->dr_stat.fs_count++ == rule->dr_drop_at;
+               __u64 count;
 
-               if (rule->dr_stat.fs_count % attr->u.drop.da_rate == 0) {
+               drop = rule->dr_stat.fs_count++ == rule->dr_drop_at;
+               count = rule->dr_stat.fs_count;
+               if (do_div(count, attr->u.drop.da_rate) == 0) {
                        rule->dr_drop_at = rule->dr_stat.fs_count +
-                                          cfs_rand() % attr->u.drop.da_rate;
+                                          prandom_u32_max(attr->u.drop.da_rate);
                        CDEBUG(D_NET, "Drop Rule %s->%s: next drop: %lu\n",
                               libcfs_nid2str(attr->fa_src),
                               libcfs_nid2str(attr->fa_dst), rule->dr_drop_at);
                }
        }
 
+drop_matched:
+
        if (drop) { /* drop this message, update counters */
+               if (hstatus)
+                       lnet_fault_match_health(hstatus,
+                               attr->u.drop.da_health_error_mask);
                lnet_fault_stat_inc(&rule->dr_stat, type);
                rule->dr_stat.u.drop.ds_dropped++;
        }
@@ -352,15 +426,17 @@ drop_rule_match(struct lnet_drop_rule *rule, lnet_nid_t src,
  * Check if message from \a src to \a dst can match any existed drop rule
  */
 bool
-lnet_drop_rule_match(lnet_hdr_t *hdr)
+lnet_drop_rule_match(struct lnet_hdr *hdr,
+                    lnet_nid_t local_nid,
+                    enum lnet_msg_hstatus *hstatus)
 {
-       struct lnet_drop_rule   *rule;
-       lnet_nid_t               src = le64_to_cpu(hdr->src_nid);
-       lnet_nid_t               dst = le64_to_cpu(hdr->dest_nid);
-       unsigned int             typ = le32_to_cpu(hdr->type);
-       unsigned int             ptl = -1;
-       bool                     drop = false;
-       int                      cpt;
+       lnet_nid_t src = le64_to_cpu(hdr->src_nid);
+       lnet_nid_t dst = le64_to_cpu(hdr->dest_nid);
+       unsigned int typ = le32_to_cpu(hdr->type);
+       struct lnet_drop_rule *rule;
+       unsigned int ptl = -1;
+       bool drop = false;
+       int cpt;
 
        /* NB: if Portal is specified, then only PUT and GET will be
         * filtered by drop rule */
@@ -371,15 +447,566 @@ lnet_drop_rule_match(lnet_hdr_t *hdr)
 
        cpt = lnet_net_lock_current();
        list_for_each_entry(rule, &the_lnet.ln_drop_rules, dr_link) {
-               drop = drop_rule_match(rule, src, dst, typ, ptl);
+               drop = drop_rule_match(rule, src, local_nid, dst, typ, ptl,
+                                      hstatus);
                if (drop)
                        break;
        }
-
        lnet_net_unlock(cpt);
+
        return drop;
 }
 
+/**
+ * LNet Delay Simulation
+ */
+/** timestamp (second) to send delayed message */
+#define msg_delay_send          msg_ev.hdr_data
+
+struct lnet_delay_rule {
+       /** link chain on the_lnet.ln_delay_rules */
+       struct list_head        dl_link;
+       /** link chain on delay_dd.dd_sched_rules */
+       struct list_head        dl_sched_link;
+       /** attributes of this rule */
+       struct lnet_fault_attr  dl_attr;
+       /** lock to protect \a below members */
+       spinlock_t              dl_lock;
+       /** refcount of delay rule */
+       atomic_t                dl_refcount;
+       /**
+        * the message sequence to delay, which means message is delayed when
+        * dl_stat.fs_count == dl_delay_at
+        */
+       unsigned long           dl_delay_at;
+       /**
+        * seconds to delay the next message, it's exclusive with dl_delay_at
+        */
+       time64_t                dl_delay_time;
+       /** baseline to caculate dl_delay_time */
+       time64_t                dl_time_base;
+       /** seconds until we send the next delayed message */
+       time64_t                dl_msg_send;
+       /** delayed message list */
+       struct list_head        dl_msg_list;
+       /** statistic of delayed messages */
+       struct lnet_fault_stat  dl_stat;
+       /** timer to wakeup delay_daemon */
+       struct timer_list       dl_timer;
+};
+
+struct delay_daemon_data {
+       /** serialise rule add/remove */
+       struct mutex            dd_mutex;
+       /** protect rules on \a dd_sched_rules */
+       spinlock_t              dd_lock;
+       /** scheduled delay rules (by timer) */
+       struct list_head        dd_sched_rules;
+       /** deamon thread sleeps at here */
+       wait_queue_head_t       dd_waitq;
+       /** controler (lctl command) wait at here */
+       wait_queue_head_t       dd_ctl_waitq;
+       /** deamon is running */
+       unsigned int            dd_running;
+       /** deamon stopped */
+       unsigned int            dd_stopped;
+};
+
+static struct delay_daemon_data        delay_dd;
+
+static void
+delay_rule_decref(struct lnet_delay_rule *rule)
+{
+       if (atomic_dec_and_test(&rule->dl_refcount)) {
+               LASSERT(list_empty(&rule->dl_sched_link));
+               LASSERT(list_empty(&rule->dl_msg_list));
+               LASSERT(list_empty(&rule->dl_link));
+
+               CFS_FREE_PTR(rule);
+       }
+}
+
+/**
+ * check source/destination NID, portal, message type and delay rate,
+ * decide whether should delay this message or not
+ */
+static bool
+delay_rule_match(struct lnet_delay_rule *rule, lnet_nid_t src,
+               lnet_nid_t dst, unsigned int type, unsigned int portal,
+               struct lnet_msg *msg)
+{
+       struct lnet_fault_attr  *attr = &rule->dl_attr;
+       bool                     delay;
+
+       if (!lnet_fault_attr_match(attr, src, LNET_NID_ANY,
+                                  dst, type, portal))
+               return false;
+
+       /* match this rule, check delay rate now */
+       spin_lock(&rule->dl_lock);
+       if (rule->dl_delay_time != 0) { /* time based delay */
+               time64_t now = ktime_get_seconds();
+
+               rule->dl_stat.fs_count++;
+               delay = now >= rule->dl_delay_time;
+               if (delay) {
+                       if (now > rule->dl_time_base)
+                               rule->dl_time_base = now;
+
+                       rule->dl_delay_time = rule->dl_time_base +
+                                             prandom_u32_max(attr->u.delay.la_interval);
+                       rule->dl_time_base += attr->u.delay.la_interval;
+
+                       CDEBUG(D_NET, "Delay Rule %s->%s: next delay : %lld\n",
+                              libcfs_nid2str(attr->fa_src),
+                              libcfs_nid2str(attr->fa_dst),
+                              rule->dl_delay_time);
+               }
+
+       } else { /* rate based delay */
+               __u64 count;
+
+               delay = rule->dl_stat.fs_count++ == rule->dl_delay_at;
+               /* generate the next random rate sequence */
+               count = rule->dl_stat.fs_count;
+               if (do_div(count, attr->u.delay.la_rate) == 0) {
+                       rule->dl_delay_at = rule->dl_stat.fs_count +
+                                           prandom_u32_max(attr->u.delay.la_rate);
+                       CDEBUG(D_NET, "Delay Rule %s->%s: next delay: %lu\n",
+                              libcfs_nid2str(attr->fa_src),
+                              libcfs_nid2str(attr->fa_dst), rule->dl_delay_at);
+               }
+       }
+
+       if (!delay) {
+               spin_unlock(&rule->dl_lock);
+               return false;
+       }
+
+       /* delay this message, update counters */
+       lnet_fault_stat_inc(&rule->dl_stat, type);
+       rule->dl_stat.u.delay.ls_delayed++;
+
+       list_add_tail(&msg->msg_list, &rule->dl_msg_list);
+       msg->msg_delay_send = ktime_get_seconds() + attr->u.delay.la_latency;
+       if (rule->dl_msg_send == -1) {
+               rule->dl_msg_send = msg->msg_delay_send;
+               mod_timer(&rule->dl_timer,
+                         jiffies + cfs_time_seconds(rule->dl_msg_send));
+       }
+
+       spin_unlock(&rule->dl_lock);
+       return true;
+}
+
+/**
+ * check if \a msg can match any Delay Rule, receiving of this message
+ * will be delayed if there is a match.
+ */
+bool
+lnet_delay_rule_match_locked(struct lnet_hdr *hdr, struct lnet_msg *msg)
+{
+       struct lnet_delay_rule  *rule;
+       lnet_nid_t               src = le64_to_cpu(hdr->src_nid);
+       lnet_nid_t               dst = le64_to_cpu(hdr->dest_nid);
+       unsigned int             typ = le32_to_cpu(hdr->type);
+       unsigned int             ptl = -1;
+
+       /* NB: called with hold of lnet_net_lock */
+
+       /* NB: if Portal is specified, then only PUT and GET will be
+        * filtered by delay rule */
+       if (typ == LNET_MSG_PUT)
+               ptl = le32_to_cpu(hdr->msg.put.ptl_index);
+       else if (typ == LNET_MSG_GET)
+               ptl = le32_to_cpu(hdr->msg.get.ptl_index);
+
+       list_for_each_entry(rule, &the_lnet.ln_delay_rules, dl_link) {
+               if (delay_rule_match(rule, src, dst, typ, ptl, msg))
+                       return true;
+       }
+
+       return false;
+}
+
+/** check out delayed messages for send */
+static void
+delayed_msg_check(struct lnet_delay_rule *rule, bool all,
+                 struct list_head *msg_list)
+{
+       struct lnet_msg *msg;
+       struct lnet_msg *tmp;
+       time64_t now = ktime_get_seconds();
+
+       if (!all && cfs_time_seconds(rule->dl_msg_send) > now)
+               return;
+
+       spin_lock(&rule->dl_lock);
+       list_for_each_entry_safe(msg, tmp, &rule->dl_msg_list, msg_list) {
+               if (!all && msg->msg_delay_send > now)
+                       break;
+
+               msg->msg_delay_send = 0;
+               list_move_tail(&msg->msg_list, msg_list);
+       }
+
+       if (list_empty(&rule->dl_msg_list)) {
+               del_timer(&rule->dl_timer);
+               rule->dl_msg_send = -1;
+
+       } else if (!list_empty(msg_list)) {
+               /* dequeued some timedout messages, update timer for the
+                * next delayed message on rule */
+               msg = list_entry(rule->dl_msg_list.next,
+                                struct lnet_msg, msg_list);
+               rule->dl_msg_send = msg->msg_delay_send;
+               mod_timer(&rule->dl_timer,
+                         jiffies + cfs_time_seconds(rule->dl_msg_send));
+       }
+       spin_unlock(&rule->dl_lock);
+}
+
+static void
+delayed_msg_process(struct list_head *msg_list, bool drop)
+{
+       struct lnet_msg *msg;
+
+       while (!list_empty(msg_list)) {
+               struct lnet_ni *ni;
+               int             cpt;
+               int             rc;
+
+               msg = list_entry(msg_list->next, struct lnet_msg, msg_list);
+               LASSERT(msg->msg_rxpeer != NULL);
+               LASSERT(msg->msg_rxni != NULL);
+
+               ni = msg->msg_rxni;
+               cpt = msg->msg_rx_cpt;
+
+               list_del_init(&msg->msg_list);
+               if (drop) {
+                       rc = -ECANCELED;
+
+               } else if (!msg->msg_routing) {
+                       rc = lnet_parse_local(ni, msg);
+                       if (rc == 0)
+                               continue;
+
+               } else {
+                       lnet_net_lock(cpt);
+                       rc = lnet_parse_forward_locked(ni, msg);
+                       lnet_net_unlock(cpt);
+
+                       switch (rc) {
+                       case LNET_CREDIT_OK:
+                               lnet_ni_recv(ni, msg->msg_private, msg, 0,
+                                            0, msg->msg_len, msg->msg_len);
+                       case LNET_CREDIT_WAIT:
+                               continue;
+                       default: /* failures */
+                               break;
+                       }
+               }
+
+               lnet_drop_message(ni, cpt, msg->msg_private, msg->msg_len,
+                                 msg->msg_type);
+               lnet_finalize(msg, rc);
+       }
+}
+
+/**
+ * Process delayed messages for scheduled rules
+ * This function can either be called by delay_rule_daemon, or by lnet_finalise
+ */
+void
+lnet_delay_rule_check(void)
+{
+       struct lnet_delay_rule *rule;
+       LIST_HEAD(msgs);
+
+       while (1) {
+               if (list_empty(&delay_dd.dd_sched_rules))
+                       break;
+
+               spin_lock_bh(&delay_dd.dd_lock);
+               if (list_empty(&delay_dd.dd_sched_rules)) {
+                       spin_unlock_bh(&delay_dd.dd_lock);
+                       break;
+               }
+
+               rule = list_entry(delay_dd.dd_sched_rules.next,
+                                 struct lnet_delay_rule, dl_sched_link);
+               list_del_init(&rule->dl_sched_link);
+               spin_unlock_bh(&delay_dd.dd_lock);
+
+               delayed_msg_check(rule, false, &msgs);
+               delay_rule_decref(rule); /* -1 for delay_dd.dd_sched_rules */
+       }
+
+       if (!list_empty(&msgs))
+               delayed_msg_process(&msgs, false);
+}
+
+/** deamon thread to handle delayed messages */
+static int
+lnet_delay_rule_daemon(void *arg)
+{
+       delay_dd.dd_running = 1;
+       wake_up(&delay_dd.dd_ctl_waitq);
+
+       while (delay_dd.dd_running) {
+               wait_event_interruptible(delay_dd.dd_waitq,
+                                        !delay_dd.dd_running ||
+                                        !list_empty(&delay_dd.dd_sched_rules));
+               lnet_delay_rule_check();
+       }
+
+       /* in case more rules have been enqueued after my last check */
+       lnet_delay_rule_check();
+       delay_dd.dd_stopped = 1;
+       wake_up(&delay_dd.dd_ctl_waitq);
+
+       return 0;
+}
+
+static void
+delay_timer_cb(cfs_timer_cb_arg_t data)
+{
+       struct lnet_delay_rule *rule = cfs_from_timer(rule, data, dl_timer);
+
+       spin_lock_bh(&delay_dd.dd_lock);
+       if (list_empty(&rule->dl_sched_link) && delay_dd.dd_running) {
+               atomic_inc(&rule->dl_refcount);
+               list_add_tail(&rule->dl_sched_link, &delay_dd.dd_sched_rules);
+               wake_up(&delay_dd.dd_waitq);
+       }
+       spin_unlock_bh(&delay_dd.dd_lock);
+}
+
+/**
+ * Add a new delay rule to LNet
+ * There is no check for duplicated delay rule, all rules will be checked for
+ * incoming message.
+ */
+int
+lnet_delay_rule_add(struct lnet_fault_attr *attr)
+{
+       struct lnet_delay_rule *rule;
+       int                     rc = 0;
+       ENTRY;
+
+       if (!((attr->u.delay.la_rate == 0) ^
+             (attr->u.delay.la_interval == 0))) {
+               CDEBUG(D_NET,
+                      "please provide either delay rate or delay interval, "
+                      "but not both at the same time %d/%d\n",
+                      attr->u.delay.la_rate, attr->u.delay.la_interval);
+               RETURN(-EINVAL);
+       }
+
+       if (attr->u.delay.la_latency == 0) {
+               CDEBUG(D_NET, "delay latency cannot be zero\n");
+               RETURN(-EINVAL);
+       }
+
+       if (lnet_fault_attr_validate(attr) != 0)
+               RETURN(-EINVAL);
+
+       CFS_ALLOC_PTR(rule);
+       if (rule == NULL)
+               RETURN(-ENOMEM);
+
+       mutex_lock(&delay_dd.dd_mutex);
+       if (!delay_dd.dd_running) {
+               struct task_struct *task;
+
+               /* NB: although LND threads will process delayed message
+                * in lnet_finalize, but there is no guarantee that LND
+                * threads will be waken up if no other message needs to
+                * be handled.
+                * Only one daemon thread, performance is not the concern
+                * of this simualation module.
+                */
+               task = kthread_run(lnet_delay_rule_daemon, NULL, "lnet_dd");
+               if (IS_ERR(task)) {
+                       rc = PTR_ERR(task);
+                       GOTO(failed, rc);
+               }
+               wait_event(delay_dd.dd_ctl_waitq, delay_dd.dd_running);
+       }
+
+       cfs_timer_setup(&rule->dl_timer, delay_timer_cb,
+                       (unsigned long)rule, 0);
+
+       spin_lock_init(&rule->dl_lock);
+       INIT_LIST_HEAD(&rule->dl_msg_list);
+       INIT_LIST_HEAD(&rule->dl_sched_link);
+
+       rule->dl_attr = *attr;
+       if (attr->u.delay.la_interval != 0) {
+               rule->dl_time_base = ktime_get_seconds() +
+                                    attr->u.delay.la_interval;
+               rule->dl_delay_time = ktime_get_seconds() +
+                                     prandom_u32_max(attr->u.delay.la_interval);
+       } else {
+               rule->dl_delay_at = prandom_u32_max(attr->u.delay.la_rate);
+       }
+
+       rule->dl_msg_send = -1;
+
+       lnet_net_lock(LNET_LOCK_EX);
+       atomic_set(&rule->dl_refcount, 1);
+       list_add(&rule->dl_link, &the_lnet.ln_delay_rules);
+       lnet_net_unlock(LNET_LOCK_EX);
+
+       CDEBUG(D_NET, "Added delay rule: src %s, dst %s, rate %d\n",
+              libcfs_nid2str(attr->fa_src), libcfs_nid2str(attr->fa_src),
+              attr->u.delay.la_rate);
+
+       mutex_unlock(&delay_dd.dd_mutex);
+       RETURN(0);
+ failed:
+       mutex_unlock(&delay_dd.dd_mutex);
+       CFS_FREE_PTR(rule);
+       return rc;
+}
+
+/**
+ * Remove matched Delay Rules from lnet, if \a shutdown is true or both \a src
+ * and \a dst are zero, all rules will be removed, otherwise only matched rules
+ * will be removed.
+ * If \a src is zero, then all rules have \a dst as destination will be remove
+ * If \a dst is zero, then all rules have \a src as source will be removed
+ *
+ * When a delay rule is removed, all delayed messages of this rule will be
+ * processed immediately.
+ */
+int
+lnet_delay_rule_del(lnet_nid_t src, lnet_nid_t dst, bool shutdown)
+{
+       struct lnet_delay_rule *rule;
+       struct lnet_delay_rule *tmp;
+       LIST_HEAD(rule_list);
+       LIST_HEAD(msg_list);
+       int n = 0;
+       bool cleanup;
+       ENTRY;
+
+       if (shutdown)
+               src = dst = 0;
+
+       mutex_lock(&delay_dd.dd_mutex);
+       lnet_net_lock(LNET_LOCK_EX);
+
+       list_for_each_entry_safe(rule, tmp, &the_lnet.ln_delay_rules, dl_link) {
+               if (rule->dl_attr.fa_src != src && src != 0)
+                       continue;
+
+               if (rule->dl_attr.fa_dst != dst && dst != 0)
+                       continue;
+
+               CDEBUG(D_NET, "Remove delay rule: src %s->dst: %s (1/%d, %d)\n",
+                      libcfs_nid2str(rule->dl_attr.fa_src),
+                      libcfs_nid2str(rule->dl_attr.fa_dst),
+                      rule->dl_attr.u.delay.la_rate,
+                      rule->dl_attr.u.delay.la_interval);
+               /* refcount is taken over by rule_list */
+               list_move(&rule->dl_link, &rule_list);
+       }
+
+       /* check if we need to shutdown delay_daemon */
+       cleanup = list_empty(&the_lnet.ln_delay_rules) &&
+                 !list_empty(&rule_list);
+       lnet_net_unlock(LNET_LOCK_EX);
+
+       list_for_each_entry_safe(rule, tmp, &rule_list, dl_link) {
+               list_del_init(&rule->dl_link);
+
+               del_timer_sync(&rule->dl_timer);
+               delayed_msg_check(rule, true, &msg_list);
+               delay_rule_decref(rule); /* -1 for the_lnet.ln_delay_rules */
+               n++;
+       }
+
+       if (cleanup) { /* no more delay rule, shutdown delay_daemon */
+               LASSERT(delay_dd.dd_running);
+               delay_dd.dd_running = 0;
+               wake_up(&delay_dd.dd_waitq);
+
+               while (!delay_dd.dd_stopped)
+                       wait_event(delay_dd.dd_ctl_waitq, delay_dd.dd_stopped);
+       }
+       mutex_unlock(&delay_dd.dd_mutex);
+
+       if (!list_empty(&msg_list))
+               delayed_msg_process(&msg_list, shutdown);
+
+       RETURN(n);
+}
+
+/**
+ * List Delay Rule at position of \a pos
+ */
+int
+lnet_delay_rule_list(int pos, struct lnet_fault_attr *attr,
+                   struct lnet_fault_stat *stat)
+{
+       struct lnet_delay_rule *rule;
+       int                     cpt;
+       int                     i = 0;
+       int                     rc = -ENOENT;
+       ENTRY;
+
+       cpt = lnet_net_lock_current();
+       list_for_each_entry(rule, &the_lnet.ln_delay_rules, dl_link) {
+               if (i++ < pos)
+                       continue;
+
+               spin_lock(&rule->dl_lock);
+               *attr = rule->dl_attr;
+               *stat = rule->dl_stat;
+               spin_unlock(&rule->dl_lock);
+               rc = 0;
+               break;
+       }
+
+       lnet_net_unlock(cpt);
+       RETURN(rc);
+}
+
+/**
+ * reset counters for all Delay Rules
+ */
+void
+lnet_delay_rule_reset(void)
+{
+       struct lnet_delay_rule *rule;
+       int                     cpt;
+       ENTRY;
+
+       cpt = lnet_net_lock_current();
+
+       list_for_each_entry(rule, &the_lnet.ln_delay_rules, dl_link) {
+               struct lnet_fault_attr *attr = &rule->dl_attr;
+
+               spin_lock(&rule->dl_lock);
+
+               memset(&rule->dl_stat, 0, sizeof(rule->dl_stat));
+               if (attr->u.delay.la_rate != 0) {
+                       rule->dl_delay_at = prandom_u32_max(attr->u.delay.la_rate);
+               } else {
+                       rule->dl_delay_time = ktime_get_seconds() +
+                                             prandom_u32_max(attr->u.delay.la_interval);
+                       rule->dl_time_base = ktime_get_seconds() +
+                                            attr->u.delay.la_interval;
+               }
+               spin_unlock(&rule->dl_lock);
+       }
+
+       lnet_net_unlock(cpt);
+       EXIT;
+}
+
 int
 lnet_fault_ctl(int opc, struct libcfs_ioctl_data *data)
 {
@@ -416,16 +1043,47 @@ lnet_fault_ctl(int opc, struct libcfs_ioctl_data *data)
                        return -EINVAL;
 
                return lnet_drop_rule_list(data->ioc_count, attr, stat);
+
+       case LNET_CTL_DELAY_ADD:
+               if (attr == NULL)
+                       return -EINVAL;
+
+               return lnet_delay_rule_add(attr);
+
+       case LNET_CTL_DELAY_DEL:
+               if (attr == NULL)
+                       return -EINVAL;
+
+               data->ioc_count = lnet_delay_rule_del(attr->fa_src,
+                                                     attr->fa_dst, false);
+               return 0;
+
+       case LNET_CTL_DELAY_RESET:
+               lnet_delay_rule_reset();
+               return 0;
+
+       case LNET_CTL_DELAY_LIST:
+               stat = (struct lnet_fault_stat *)data->ioc_inlbuf2;
+               if (attr == NULL || stat == NULL)
+                       return -EINVAL;
+
+               return lnet_delay_rule_list(data->ioc_count, attr, stat);
        }
 }
 
 int
 lnet_fault_init(void)
 {
-       CLASSERT(LNET_PUT_BIT == 1 << LNET_MSG_PUT);
-       CLASSERT(LNET_ACK_BIT == 1 << LNET_MSG_ACK);
-       CLASSERT(LNET_GET_BIT == 1 << LNET_MSG_GET);
-       CLASSERT(LNET_REPLY_BIT == 1 << LNET_MSG_REPLY);
+       BUILD_BUG_ON(LNET_PUT_BIT != 1 << LNET_MSG_PUT);
+       BUILD_BUG_ON(LNET_ACK_BIT != 1 << LNET_MSG_ACK);
+       BUILD_BUG_ON(LNET_GET_BIT != 1 << LNET_MSG_GET);
+       BUILD_BUG_ON(LNET_REPLY_BIT != 1 << LNET_MSG_REPLY);
+
+       mutex_init(&delay_dd.dd_mutex);
+       spin_lock_init(&delay_dd.dd_lock);
+       init_waitqueue_head(&delay_dd.dd_waitq);
+       init_waitqueue_head(&delay_dd.dd_ctl_waitq);
+       INIT_LIST_HEAD(&delay_dd.dd_sched_rules);
 
        return 0;
 }
@@ -434,6 +1092,9 @@ void
 lnet_fault_fini(void)
 {
        lnet_drop_rule_del(0, 0);
+       lnet_delay_rule_del(0, 0, true);
 
        LASSERT(list_empty(&the_lnet.ln_drop_rules));
+       LASSERT(list_empty(&the_lnet.ln_delay_rules));
+       LASSERT(list_empty(&delay_dd.dd_sched_rules));
 }