X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=blobdiff_plain;f=lnet%2Flnet%2Flib-md.c;h=c13993466e04193110387d1e0e26ee6f7589831f;hp=6e64ad254a61b3a42f3949c67545de458c636acf;hb=faeb94fe81e4646b3121c263521d30e0e83fa71f;hpb=5e1e6a6756d3b4ca19a0d7e0defcf974dbfed13c diff --git a/lnet/lnet/lib-md.c b/lnet/lnet/lib-md.c index 6e64ad2..c139934 100644 --- a/lnet/lnet/lib-md.c +++ b/lnet/lnet/lib-md.c @@ -1,6 +1,4 @@ -/* -*- mode: c; c-basic-offset: 8; indent-tabs-mode: nil; -*- - * vim:expandtab:shiftwidth=8:tabstop=8: - * +/* * GPL HEADER START * * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. @@ -26,8 +24,10 @@ * GPL HEADER END */ /* - * Copyright 2008 Sun Microsystems, Inc. All rights reserved + * Copyright (c) 2003, 2010, Oracle and/or its affiliates. All rights reserved. * Use is subject to license terms. + * + * Copyright (c) 2012, 2013, Intel Corporation. */ /* * This file is part of Lustre, http://www.lustre.org/ @@ -42,7 +42,7 @@ #include -/* must be called with LNET_LOCK held */ +/* must be called with lnet_res_lock held */ void lnet_md_unlink(lnet_libmd_t *md) { @@ -55,14 +55,14 @@ lnet_md_unlink(lnet_libmd_t *md) /* Disassociate from ME (if any), and unlink it if it was created * with LNET_UNLINK */ if (me != NULL) { - md->md_me = NULL; - me->me_md = NULL; + /* detach MD from portal */ + lnet_ptl_detach_md(me, md); if (me->me_unlink == LNET_UNLINK) lnet_me_unlink(me); } /* ensure all future handle lookups fail */ - lnet_invalidate_handle(&md->md_lh); + lnet_res_lh_invalidate(&md->md_lh); } if (md->md_refcount != 0) { @@ -73,51 +73,31 @@ lnet_md_unlink(lnet_libmd_t *md) CDEBUG(D_NET, "Unlinking md %p\n", md); if (md->md_eq != NULL) { - md->md_eq->eq_refcount--; - LASSERT (md->md_eq->eq_refcount >= 0); - } + int cpt = lnet_cpt_of_cookie(md->md_lh.lh_cookie); + + LASSERT(*md->md_eq->eq_refs[cpt] > 0); + (*md->md_eq->eq_refs[cpt])--; + } - LASSERT (!list_empty(&md->md_list)); - list_del_init (&md->md_list); - lnet_md_free(md); + LASSERT(!list_empty(&md->md_list)); + list_del_init(&md->md_list); + lnet_md_free(md); } -/* must be called with LNET_LOCK held */ static int -lib_md_build(lnet_libmd_t *lmd, lnet_md_t *umd, int unlink) +lnet_md_build(lnet_libmd_t *lmd, lnet_md_t *umd, int unlink) { - lnet_eq_t *eq = NULL; int i; unsigned int niov; int total_length = 0; - /* NB we are passed an allocated, but uninitialised/active md. - * if we return success, caller may lnet_md_unlink() it. - * otherwise caller may only lnet_md_free() it. - */ - - if (!LNetHandleIsInvalid (umd->eq_handle)) { - eq = lnet_handle2eq(&umd->eq_handle); - if (eq == NULL) - return -ENOENT; - } - - /* This implementation doesn't know how to create START events or - * disable END events. Best to LASSERT our caller is compliant so - * we find out quickly... */ - /* TODO - reevaluate what should be here in light of - * the removal of the start and end events - * maybe there we shouldn't even allow LNET_EQ_NONE!) - LASSERT (eq == NULL); - */ - lmd->md_me = NULL; lmd->md_start = umd->start; lmd->md_offset = 0; lmd->md_max_size = umd->max_size; lmd->md_options = umd->options; lmd->md_user_ptr = umd->user_ptr; - lmd->md_eq = eq; + lmd->md_eq = NULL; lmd->md_threshold = umd->threshold; lmd->md_refcount = 0; lmd->md_flags = (unlink == LNET_UNLINK) ? LNET_MD_FLAG_AUTO_UNLINK : 0; @@ -147,9 +127,6 @@ lib_md_build(lnet_libmd_t *lmd, lnet_md_t *umd, int unlink) return -EINVAL; } else if ((umd->options & LNET_MD_KIOV) != 0) { -#ifndef __KERNEL__ - return -EINVAL; -#else lmd->md_niov = niov = umd->length; memcpy(lmd->md_iov.kiov, umd->start, niov * sizeof (lmd->md_iov.kiov[0])); @@ -157,7 +134,7 @@ lib_md_build(lnet_libmd_t *lmd, lnet_md_t *umd, int unlink) for (i = 0; i < (int)niov; i++) { /* We take the page pointer on trust */ if (lmd->md_iov.kiov[i].kiov_offset + - lmd->md_iov.kiov[i].kiov_len > CFS_PAGE_SIZE ) + lmd->md_iov.kiov[i].kiov_len > PAGE_CACHE_SIZE) return -EINVAL; /* invalid length */ total_length += lmd->md_iov.kiov[i].kiov_len; @@ -169,7 +146,6 @@ lib_md_build(lnet_libmd_t *lmd, lnet_md_t *umd, int unlink) (umd->max_size < 0 || umd->max_size > total_length)) // illegal max_size return -EINVAL; -#endif } else { /* contiguous */ lmd->md_length = umd->length; lmd->md_niov = niov = 1; @@ -182,18 +158,45 @@ lib_md_build(lnet_libmd_t *lmd, lnet_md_t *umd, int unlink) return -EINVAL; } - if (eq != NULL) - eq->eq_refcount++; - - /* It's good; let handle2md succeed and add to active mds */ - lnet_initialise_handle (&lmd->md_lh, LNET_COOKIE_TYPE_MD); - LASSERT (list_empty(&lmd->md_list)); - list_add (&lmd->md_list, &the_lnet.ln_active_mds); + return 0; +} - return 0; +/* must be called with resource lock held */ +static int +lnet_md_link(lnet_libmd_t *md, lnet_handle_eq_t eq_handle, int cpt) +{ + struct lnet_res_container *container = the_lnet.ln_md_containers[cpt]; + + /* NB we are passed an allocated, but inactive md. + * if we return success, caller may lnet_md_unlink() it. + * otherwise caller may only lnet_md_free() it. + */ + /* This implementation doesn't know how to create START events or + * disable END events. Best to LASSERT our caller is compliant so + * we find out quickly... */ + /* TODO - reevaluate what should be here in light of + * the removal of the start and end events + * maybe there we shouldn't even allow LNET_EQ_NONE!) + * LASSERT (eq == NULL); + */ + if (!LNetHandleIsInvalid(eq_handle)) { + md->md_eq = lnet_handle2eq(&eq_handle); + + if (md->md_eq == NULL) + return -ENOENT; + + (*md->md_eq->eq_refs[cpt])++; + } + + lnet_res_lh_initialize(container, &md->md_lh); + + LASSERT(list_empty(&md->md_list)); + list_add(&md->md_list, &container->rec_active); + + return 0; } -/* must be called with LNET_LOCK held */ +/* must be called with lnet_res_lock held */ void lnet_md_deconstruct(lnet_libmd_t *lmd, lnet_md_t *umd) { @@ -212,17 +215,18 @@ lnet_md_deconstruct(lnet_libmd_t *lmd, lnet_md_t *umd) lnet_eq2handle(&umd->eq_handle, lmd->md_eq); } -int +static int lnet_md_validate(lnet_md_t *umd) { - if (umd->start == NULL) { - CERROR("MD start pointer can not be NULL\n"); + if (umd->start == NULL && umd->length != 0) { + CERROR("MD start pointer can not be NULL with length %u\n", + umd->length); return -EINVAL; } if ((umd->options & (LNET_MD_KIOV | LNET_MD_IOVEC)) != 0 && umd->length > LNET_MAX_IOV) { - CERROR("Invalid option: too many fragments %d, %d max\n", + CERROR("Invalid option: too many fragments %u, %d max\n", umd->length, LNET_MAX_IOV); return -EINVAL; } @@ -230,15 +234,42 @@ lnet_md_validate(lnet_md_t *umd) return 0; } +/** + * Create a memory descriptor and attach it to a ME + * + * \param meh A handle for a ME to associate the new MD with. + * \param umd Provides initial values for the user-visible parts of a MD. + * Other than its use for initialization, there is no linkage between this + * structure and the MD maintained by the LNet. + * \param unlink A flag to indicate whether the MD is automatically unlinked + * when it becomes inactive, either because the operation threshold drops to + * zero or because the available memory becomes less than \a umd.max_size. + * (Note that the check for unlinking a MD only occurs after the completion + * of a successful operation on the MD.) The value LNET_UNLINK enables auto + * unlinking; the value LNET_RETAIN disables it. + * \param handle On successful returns, a handle to the newly created MD is + * saved here. This handle can be used later in LNetMDUnlink(). + * + * \retval 0 On success. + * \retval -EINVAL If \a umd is not valid. + * \retval -ENOMEM If new MD cannot be allocated. + * \retval -ENOENT Either \a meh or \a umd.eq_handle does not point to a + * valid object. Note that it's OK to supply a NULL \a umd.eq_handle by + * calling LNetInvalidateHandle() on it. + * \retval -EBUSY If the ME pointed to by \a meh is already associated with + * a MD. + */ int LNetMDAttach(lnet_handle_me_t meh, lnet_md_t umd, - lnet_unlink_t unlink, lnet_handle_md_t *handle) + lnet_unlink_t unlink, lnet_handle_md_t *handle) { - lnet_me_t *me; - lnet_libmd_t *md; - int rc; + struct list_head matches = LIST_HEAD_INIT(matches); + struct list_head drops = LIST_HEAD_INIT(drops); + struct lnet_me *me; + struct lnet_libmd *md; + int cpt; + int rc; - LASSERT (the_lnet.ln_init); LASSERT (the_lnet.ln_refcount > 0); if (lnet_md_validate(&umd) != 0) @@ -253,44 +284,68 @@ LNetMDAttach(lnet_handle_me_t meh, lnet_md_t umd, if (md == NULL) return -ENOMEM; - LNET_LOCK(); + rc = lnet_md_build(md, &umd, unlink); + cpt = lnet_cpt_of_cookie(meh.cookie); - me = lnet_handle2me(&meh); - if (me == NULL) { - rc = -ENOENT; - } else if (me->me_md != NULL) { - rc = -EBUSY; - } else { - rc = lib_md_build(md, &umd, unlink); - if (rc == 0) { - the_lnet.ln_portals[me->me_portal].ptl_ml_version++; + lnet_res_lock(cpt); + if (rc != 0) + goto failed; - me->me_md = md; - md->md_me = me; + me = lnet_handle2me(&meh); + if (me == NULL) + rc = -ENOENT; + else if (me->me_md != NULL) + rc = -EBUSY; + else + rc = lnet_md_link(md, umd.eq_handle, cpt); - lnet_md2handle(handle, md); + if (rc != 0) + goto failed; - /* check if this MD matches any blocked msgs */ - lnet_match_blocked_msg(md); /* expects LNET_LOCK held */ + /* attach this MD to portal of ME and check if it matches any + * blocked msgs on this portal */ + lnet_ptl_attach_md(me, md, &matches, &drops); - LNET_UNLOCK(); - return (0); - } - } + lnet_md2handle(handle, md); + + lnet_res_unlock(cpt); - lnet_md_free (md); + lnet_drop_delayed_msg_list(&drops, "Bad match"); + lnet_recv_delayed_msg_list(&matches); - LNET_UNLOCK(); - return (rc); + return 0; + + failed: + lnet_md_free(md); + + lnet_res_unlock(cpt); + return rc; } +EXPORT_SYMBOL(LNetMDAttach); +/** + * Create a "free floating" memory descriptor - a MD that is not associated + * with a ME. Such MDs are usually used in LNetPut() and LNetGet() operations. + * + * \param umd,unlink See the discussion for LNetMDAttach(). + * \param handle On successful returns, a handle to the newly created MD is + * saved here. This handle can be used later in LNetMDUnlink(), LNetPut(), + * and LNetGet() operations. + * + * \retval 0 On success. + * \retval -EINVAL If \a umd is not valid. + * \retval -ENOMEM If new MD cannot be allocated. + * \retval -ENOENT \a umd.eq_handle does not point to a valid EQ. Note that + * it's OK to supply a NULL \a umd.eq_handle by calling + * LNetInvalidateHandle() on it. + */ int LNetMDBind(lnet_md_t umd, lnet_unlink_t unlink, lnet_handle_md_t *handle) { - lnet_libmd_t *md; - int rc; + lnet_libmd_t *md; + int cpt; + int rc; - LASSERT (the_lnet.ln_init); LASSERT (the_lnet.ln_refcount > 0); if (lnet_md_validate(&umd) != 0) @@ -305,52 +360,89 @@ LNetMDBind(lnet_md_t umd, lnet_unlink_t unlink, lnet_handle_md_t *handle) if (md == NULL) return -ENOMEM; - LNET_LOCK(); + rc = lnet_md_build(md, &umd, unlink); - rc = lib_md_build(md, &umd, unlink); + cpt = lnet_res_lock_current(); + if (rc != 0) + goto failed; - if (rc == 0) { - lnet_md2handle(handle, md); + rc = lnet_md_link(md, umd.eq_handle, cpt); + if (rc != 0) + goto failed; - LNET_UNLOCK(); - return (0); - } + lnet_md2handle(handle, md); - lnet_md_free (md); + lnet_res_unlock(cpt); + return 0; - LNET_UNLOCK(); - return (rc); + failed: + lnet_md_free(md); + + lnet_res_unlock(cpt); + return rc; } +EXPORT_SYMBOL(LNetMDBind); +/** + * Unlink the memory descriptor from any ME it may be linked to and release + * the internal resources associated with it. As a result, active messages + * associated with the MD may get aborted. + * + * This function does not free the memory region associated with the MD; + * i.e., the memory the user allocated for this MD. If the ME associated with + * this MD is not NULL and was created with auto unlink enabled, the ME is + * unlinked as well (see LNetMEAttach()). + * + * Explicitly unlinking a MD via this function call has the same behavior as + * a MD that has been automatically unlinked, except that no LNET_EVENT_UNLINK + * is generated in the latter case. + * + * An unlinked event can be reported in two ways: + * - If there's no pending operations on the MD, it's unlinked immediately + * and an LNET_EVENT_UNLINK event is logged before this function returns. + * - Otherwise, the MD is only marked for deletion when this function + * returns, and the unlinked event will be piggybacked on the event of + * the completion of the last operation by setting the unlinked field of + * the event. No dedicated LNET_EVENT_UNLINK event is generated. + * + * Note that in both cases the unlinked field of the event is always set; no + * more event will happen on the MD after such an event is logged. + * + * \param mdh A handle for the MD to be unlinked. + * + * \retval 0 On success. + * \retval -ENOENT If \a mdh does not point to a valid MD object. + */ int LNetMDUnlink (lnet_handle_md_t mdh) { - lnet_event_t ev; - lnet_libmd_t *md; + lnet_event_t ev; + lnet_libmd_t *md; + int cpt; - LASSERT (the_lnet.ln_init); - LASSERT (the_lnet.ln_refcount > 0); + LASSERT(the_lnet.ln_refcount > 0); - LNET_LOCK(); + cpt = lnet_cpt_of_cookie(mdh.cookie); + lnet_res_lock(cpt); - md = lnet_handle2md(&mdh); - if (md == NULL) { - LNET_UNLOCK(); + md = lnet_handle2md(&mdh); + if (md == NULL) { + lnet_res_unlock(cpt); return -ENOENT; } - /* If the MD is busy, lnet_md_unlink just marks it for deletion, and - * when the NAL is done, the completion event flags that the MD was - * unlinked. Otherwise, we enqueue an event now... */ - - if (md->md_eq != NULL && - md->md_refcount == 0) { - lnet_build_unlink_event(md, &ev); - lnet_enq_event_locked(md->md_eq, &ev); - } + md->md_flags |= LNET_MD_FLAG_ABORTED; + /* If the MD is busy, lnet_md_unlink just marks it for deletion, and + * when the LND is done, the completion event flags that the MD was + * unlinked. Otherwise, we enqueue an event now... */ + if (md->md_eq != NULL && md->md_refcount == 0) { + lnet_build_unlink_event(md, &ev); + lnet_eq_enqueue_event(md->md_eq, &ev); + } lnet_md_unlink(md); - LNET_UNLOCK(); - return 0; + lnet_res_unlock(cpt); + return 0; } +EXPORT_SYMBOL(LNetMDUnlink);