/*
* Copyright (c) 2003, 2010, Oracle and/or its affiliates. All rights reserved.
* Use is subject to license terms.
+ *
+ * Copyright (c) 2012, Intel Corporation.
*/
/*
* This file is part of Lustre, http://www.lustre.org/
#include <lnet/lib-lnet.h>
-/* must be called with LNET_LOCK held */
+/* must be called with lnet_res_lock held */
void
lnet_md_unlink(lnet_libmd_t *md)
{
CDEBUG(D_NET, "Unlinking md %p\n", md);
if (md->md_eq != NULL) {
- md->md_eq->eq_refcount--;
- LASSERT (md->md_eq->eq_refcount >= 0);
- }
+ int cpt = lnet_cpt_of_cookie(md->md_lh.lh_cookie);
+
+ LASSERT(*md->md_eq->eq_refs[cpt] > 0);
+ (*md->md_eq->eq_refs[cpt])--;
+ }
- LASSERT (!cfs_list_empty(&md->md_list));
- cfs_list_del_init (&md->md_list);
+ LASSERT(!cfs_list_empty(&md->md_list));
+ cfs_list_del_init(&md->md_list);
lnet_md_free_locked(md);
}
for (i = 0; i < (int)niov; i++) {
/* We take the page pointer on trust */
if (lmd->md_iov.kiov[i].kiov_offset +
- lmd->md_iov.kiov[i].kiov_len > CFS_PAGE_SIZE )
+ lmd->md_iov.kiov[i].kiov_len > PAGE_CACHE_SIZE)
return -EINVAL; /* invalid length */
total_length += lmd->md_iov.kiov[i].kiov_len;
/* must be called with resource lock held */
static int
-lnet_md_link(lnet_libmd_t *md, lnet_handle_eq_t eq_handle)
+lnet_md_link(lnet_libmd_t *md, lnet_handle_eq_t eq_handle, int cpt)
{
- struct lnet_res_container *container = &the_lnet.ln_md_container;
+ struct lnet_res_container *container = the_lnet.ln_md_containers[cpt];
/* NB we are passed an allocated, but inactive md.
* if we return success, caller may lnet_md_unlink() it.
if (md->md_eq == NULL)
return -ENOENT;
- md->md_eq->eq_refcount++;
+ (*md->md_eq->eq_refs[cpt])++;
}
lnet_res_lh_initialize(container, &md->md_lh);
return 0;
}
-/* must be called with LNET_LOCK held */
+/* must be called with lnet_res_lock held */
void
lnet_md_deconstruct(lnet_libmd_t *lmd, lnet_md_t *umd)
{
*/
int
LNetMDAttach(lnet_handle_me_t meh, lnet_md_t umd,
- lnet_unlink_t unlink, lnet_handle_md_t *handle)
+ lnet_unlink_t unlink, lnet_handle_md_t *handle)
{
- CFS_LIST_HEAD (matches);
- CFS_LIST_HEAD (drops);
- lnet_me_t *me;
- lnet_libmd_t *md;
- int rc;
+ CFS_LIST_HEAD (matches);
+ CFS_LIST_HEAD (drops);
+ struct lnet_me *me;
+ struct lnet_libmd *md;
+ int cpt;
+ int rc;
LASSERT (the_lnet.ln_init);
LASSERT (the_lnet.ln_refcount > 0);
return -ENOMEM;
rc = lnet_md_build(md, &umd, unlink);
+ cpt = lnet_cpt_of_cookie(meh.cookie);
- LNET_LOCK();
+ lnet_res_lock(cpt);
if (rc != 0)
goto failed;
- me = lnet_handle2me(&meh);
+ me = lnet_handle2me(&meh);
if (me == NULL)
rc = -ENOENT;
else if (me->me_md != NULL)
- rc = -EBUSY;
+ rc = -EBUSY;
else
- rc = lnet_md_link(md, umd.eq_handle);
+ rc = lnet_md_link(md, umd.eq_handle, cpt);
if (rc != 0)
goto failed;
lnet_md2handle(handle, md);
- LNET_UNLOCK();
+ lnet_res_unlock(cpt);
lnet_drop_delayed_msg_list(&drops, "Bad match");
lnet_recv_delayed_msg_list(&matches);
failed:
lnet_md_free_locked(md);
- LNET_UNLOCK();
+ lnet_res_unlock(cpt);
return rc;
}
+EXPORT_SYMBOL(LNetMDAttach);
/**
* Create a "free floating" memory descriptor - a MD that is not associated
int
LNetMDBind(lnet_md_t umd, lnet_unlink_t unlink, lnet_handle_md_t *handle)
{
- lnet_libmd_t *md;
- int rc;
+ lnet_libmd_t *md;
+ int cpt;
+ int rc;
LASSERT (the_lnet.ln_init);
LASSERT (the_lnet.ln_refcount > 0);
rc = lnet_md_build(md, &umd, unlink);
- LNET_LOCK();
+ cpt = lnet_res_lock_current();
if (rc != 0)
goto failed;
- rc = lnet_md_link(md, umd.eq_handle);
+ rc = lnet_md_link(md, umd.eq_handle, cpt);
if (rc != 0)
goto failed;
lnet_md2handle(handle, md);
- LNET_UNLOCK();
+ lnet_res_unlock(cpt);
return 0;
failed:
lnet_md_free_locked(md);
- LNET_UNLOCK();
+ lnet_res_unlock(cpt);
return rc;
}
+EXPORT_SYMBOL(LNetMDBind);
/**
* Unlink the memory descriptor from any ME it may be linked to and release
- * the internal resources associated with it.
+ * the internal resources associated with it. As a result, active messages
+ * associated with the MD may get aborted.
*
* This function does not free the memory region associated with the MD;
* i.e., the memory the user allocated for this MD. If the ME associated with
int
LNetMDUnlink (lnet_handle_md_t mdh)
{
- lnet_event_t ev;
- lnet_libmd_t *md;
+ lnet_event_t ev;
+ lnet_libmd_t *md;
+ int cpt;
- LASSERT (the_lnet.ln_init);
- LASSERT (the_lnet.ln_refcount > 0);
+ LASSERT(the_lnet.ln_init);
+ LASSERT(the_lnet.ln_refcount > 0);
- LNET_LOCK();
+ cpt = lnet_cpt_of_cookie(mdh.cookie);
+ lnet_res_lock(cpt);
- md = lnet_handle2md(&mdh);
- if (md == NULL) {
- LNET_UNLOCK();
+ md = lnet_handle2md(&mdh);
+ if (md == NULL) {
+ lnet_res_unlock(cpt);
return -ENOENT;
}
- /* If the MD is busy, lnet_md_unlink just marks it for deletion, and
- * when the NAL is done, the completion event flags that the MD was
- * unlinked. Otherwise, we enqueue an event now... */
-
- if (md->md_eq != NULL &&
- md->md_refcount == 0) {
- lnet_build_unlink_event(md, &ev);
+ md->md_flags |= LNET_MD_FLAG_ABORTED;
+ /* If the MD is busy, lnet_md_unlink just marks it for deletion, and
+ * when the LND is done, the completion event flags that the MD was
+ * unlinked. Otherwise, we enqueue an event now... */
+ if (md->md_eq != NULL && md->md_refcount == 0) {
+ lnet_build_unlink_event(md, &ev);
lnet_eq_enqueue_event(md->md_eq, &ev);
- }
+ }
lnet_md_unlink(md);
- LNET_UNLOCK();
- return 0;
+ lnet_res_unlock(cpt);
+ return 0;
}
+EXPORT_SYMBOL(LNetMDUnlink);