X-Git-Url: https://git.whamcloud.com/?a=blobdiff_plain;f=lnet%2Fklnds%2Fptllnd%2Fptllnd.c;h=791f5c2b60d6b669106f7eb7a6c8688cab0c267b;hb=98060d83459ba10409f295898f0ec917f938b4d3;hp=279db5a899cb66fec9812c7e2ba4cd8c8a5313a6;hpb=e3a7c58aebafce40323db54bf6056029e5af4a70;p=fs%2Flustre-release.git diff --git a/lnet/klnds/ptllnd/ptllnd.c b/lnet/klnds/ptllnd/ptllnd.c index 279db5a..791f5c2 100644 --- a/lnet/klnds/ptllnd/ptllnd.c +++ b/lnet/klnds/ptllnd/ptllnd.c @@ -26,6 +26,8 @@ /* * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved. * Use is subject to license terms. + * + * Copyright (c) 2012, Intel Corporation. */ /* * This file is part of Lustre, http://www.lustre.org/ @@ -59,12 +61,12 @@ kptllnd_ptlid2str(ptl_process_id_t id) unsigned long flags; char *str; - - cfs_spin_lock_irqsave(&kptllnd_data.kptl_ptlid2str_lock, flags); + + spin_lock_irqsave(&kptllnd_data.kptl_ptlid2str_lock, flags); str = strs[idx++]; if (idx >= sizeof(strs)/sizeof(strs[0])) idx = 0; - cfs_spin_unlock_irqrestore(&kptllnd_data.kptl_ptlid2str_lock, flags); + spin_unlock_irqrestore(&kptllnd_data.kptl_ptlid2str_lock, flags); snprintf(str, sizeof(strs[0]), FMT_PTLID, id.pid, id.nid); return str; @@ -491,10 +493,10 @@ kptllnd_query (lnet_ni_t *ni, lnet_nid_t nid, cfs_time_t *when) if (kptllnd_find_target(net, id, &peer) != 0) return; - cfs_spin_lock_irqsave(&peer->peer_lock, flags); + spin_lock_irqsave(&peer->peer_lock, flags); if (peer->peer_last_alive != 0) *when = peer->peer_last_alive; - cfs_spin_unlock_irqrestore(&peer->peer_lock, flags); + spin_unlock_irqrestore(&peer->peer_lock, flags); kptllnd_peer_decref(peer); return; } @@ -507,9 +509,9 @@ kptllnd_base_shutdown (void) unsigned long flags; lnet_process_id_t process_id; - cfs_read_lock(&kptllnd_data.kptl_net_rw_lock); + read_lock(&kptllnd_data.kptl_net_rw_lock); LASSERT (cfs_list_empty(&kptllnd_data.kptl_nets)); - cfs_read_unlock(&kptllnd_data.kptl_net_rw_lock); + read_unlock(&kptllnd_data.kptl_net_rw_lock); switch (kptllnd_data.kptl_init) { default: @@ -523,11 +525,11 @@ kptllnd_base_shutdown (void) LASSERT (cfs_list_empty(&kptllnd_data.kptl_sched_rxbq)); /* lock to interleave cleanly with peer birth/death */ - cfs_write_lock_irqsave(&kptllnd_data.kptl_peer_rw_lock, flags); + write_lock_irqsave(&kptllnd_data.kptl_peer_rw_lock, flags); LASSERT (kptllnd_data.kptl_shutdown == 0); kptllnd_data.kptl_shutdown = 1; /* phase 1 == destroy peers */ /* no new peers possible now */ - cfs_write_unlock_irqrestore(&kptllnd_data.kptl_peer_rw_lock, + write_unlock_irqrestore(&kptllnd_data.kptl_peer_rw_lock, flags); /* nuke all existing peers */ @@ -535,7 +537,7 @@ kptllnd_base_shutdown (void) process_id.pid = LNET_PID_ANY; kptllnd_peer_del(process_id); - cfs_read_lock_irqsave(&kptllnd_data.kptl_peer_rw_lock, flags); + read_lock_irqsave(&kptllnd_data.kptl_peer_rw_lock, flags); LASSERT (kptllnd_data.kptl_n_active_peers == 0); @@ -546,12 +548,12 @@ kptllnd_base_shutdown (void) "Waiting for %d peers to terminate\n", kptllnd_data.kptl_npeers); - cfs_read_unlock_irqrestore(&kptllnd_data.kptl_peer_rw_lock, + read_unlock_irqrestore(&kptllnd_data.kptl_peer_rw_lock, flags); cfs_pause(cfs_time_seconds(1)); - cfs_read_lock_irqsave(&kptllnd_data.kptl_peer_rw_lock, + read_lock_irqsave(&kptllnd_data.kptl_peer_rw_lock, flags); } @@ -561,7 +563,7 @@ kptllnd_base_shutdown (void) for (i = 0; i < kptllnd_data.kptl_peer_hash_size; i++) LASSERT (cfs_list_empty (&kptllnd_data.kptl_peers[i])); - cfs_read_unlock_irqrestore(&kptllnd_data.kptl_peer_rw_lock, + read_unlock_irqrestore(&kptllnd_data.kptl_peer_rw_lock, flags); CDEBUG(D_NET, "All peers deleted\n"); @@ -672,21 +674,21 @@ kptllnd_base_startup (void) kptllnd_data.kptl_eqh = PTL_INVALID_HANDLE; kptllnd_data.kptl_nih = PTL_INVALID_HANDLE; - cfs_rwlock_init(&kptllnd_data.kptl_net_rw_lock); + rwlock_init(&kptllnd_data.kptl_net_rw_lock); CFS_INIT_LIST_HEAD(&kptllnd_data.kptl_nets); /* Setup the sched locks/lists/waitq */ - cfs_spin_lock_init(&kptllnd_data.kptl_sched_lock); + spin_lock_init(&kptllnd_data.kptl_sched_lock); cfs_waitq_init(&kptllnd_data.kptl_sched_waitq); CFS_INIT_LIST_HEAD(&kptllnd_data.kptl_sched_txq); CFS_INIT_LIST_HEAD(&kptllnd_data.kptl_sched_rxq); CFS_INIT_LIST_HEAD(&kptllnd_data.kptl_sched_rxbq); /* Init kptl_ptlid2str_lock before any call to kptllnd_ptlid2str */ - cfs_spin_lock_init(&kptllnd_data.kptl_ptlid2str_lock); + spin_lock_init(&kptllnd_data.kptl_ptlid2str_lock); /* Setup the tx locks/lists */ - cfs_spin_lock_init(&kptllnd_data.kptl_tx_lock); + spin_lock_init(&kptllnd_data.kptl_tx_lock); CFS_INIT_LIST_HEAD(&kptllnd_data.kptl_idle_txs); cfs_atomic_set(&kptllnd_data.kptl_ntx, 0); @@ -771,9 +773,8 @@ kptllnd_base_startup (void) kptllnd_data.kptl_nak_msg->ptlm_srcpid = the_lnet.ln_pid; kptllnd_data.kptl_nak_msg->ptlm_srcstamp = kptllnd_data.kptl_incarnation; - cfs_rwlock_init(&kptllnd_data.kptl_peer_rw_lock); + rwlock_init(&kptllnd_data.kptl_peer_rw_lock); cfs_waitq_init(&kptllnd_data.kptl_watchdog_waitq); - cfs_atomic_set(&kptllnd_data.kptl_needs_ptltrace, 0); CFS_INIT_LIST_HEAD(&kptllnd_data.kptl_closing_peers); CFS_INIT_LIST_HEAD(&kptllnd_data.kptl_zombie_peers); @@ -906,9 +907,9 @@ kptllnd_startup (lnet_ni_t *ni) kptllnd_data.kptl_nak_msg->ptlm_srcnid = ni->ni_nid; cfs_atomic_set(&net->net_refcount, 1); - cfs_write_lock(&kptllnd_data.kptl_net_rw_lock); + write_lock(&kptllnd_data.kptl_net_rw_lock); cfs_list_add_tail(&net->net_list, &kptllnd_data.kptl_nets); - cfs_write_unlock(&kptllnd_data.kptl_net_rw_lock); + write_unlock(&kptllnd_data.kptl_net_rw_lock); return 0; failed: @@ -938,15 +939,15 @@ kptllnd_shutdown (lnet_ni_t *ni) ni->ni_data = NULL; net->net_ni = NULL; - cfs_write_lock(&kptllnd_data.kptl_net_rw_lock); + write_lock(&kptllnd_data.kptl_net_rw_lock); kptllnd_net_decref(net); cfs_list_del_init(&net->net_list); - cfs_write_unlock(&kptllnd_data.kptl_net_rw_lock); + write_unlock(&kptllnd_data.kptl_net_rw_lock); /* Can't nuke peers here - they are shared among all NIs */ - cfs_write_lock_irqsave(&kptllnd_data.kptl_peer_rw_lock, flags); + write_lock_irqsave(&kptllnd_data.kptl_peer_rw_lock, flags); net->net_shutdown = 1; /* Order with peer creation */ - cfs_write_unlock_irqrestore(&kptllnd_data.kptl_peer_rw_lock, flags); + write_unlock_irqrestore(&kptllnd_data.kptl_peer_rw_lock, flags); i = 2; while (cfs_atomic_read(&net->net_refcount) != 0) {