portals spec.
/* save transaction info for later finalize and cleanup */
PORTAL_ALLOC(ktx, (sizeof(kscimacnal_tx_t)));
if (!ktx) {
- return PTL_NOSPACE;
+ return PTL_NO_SPACE;
}
ktx->ktx_nmapped = 0; /* Start with no mapped pages :) */
kscimacnal_txrelease, ktx);
if (!msg) {
PORTAL_FREE(ktx, (sizeof(kscimacnal_tx_t)));
- return PTL_NOSPACE;
+ return PTL_NO_SPACE;
}
mac_put_mblk(msg, sizeof(ptl_hdr_t));
lastblk=msg;
if(!newblk) {
mac_free_msg(msg);
PORTAL_FREE(ktx, (sizeof(kscimacnal_tx_t)));
- return PTL_NOSPACE;
+ return PTL_NO_SPACE;
}
mac_put_mblk(newblk, nob);
mac_link_mblk(lastblk, newblk);
unsigned long flags;
if (!ptl_init)
- RETURN(PTL_NOINIT);
+ RETURN(PTL_NO_INIT);
if (neq_in < 1)
- RETURN(PTL_INV_EQ);
+ RETURN(PTL_EQ_INVALID);
nal = ptl_hndl2nal(&eventqs_in[0]);
if (nal == NULL)
- RETURN(PTL_INV_EQ);
+ RETURN(PTL_EQ_INVALID);
nal->lock(nal, &flags);
if (i > 0 &&
ptl_hndl2nal(&eventqs_in[i]) != nal) {
nal->unlock(nal, &flags);
- RETURN (PTL_INV_EQ);
+ RETURN (PTL_EQ_INVALID);
}
/* size must be a power of 2 to handle a wrapped sequence # */
if (ptl_init)
return PTL_OK;
+ LASSERT(!strcmp(ptl_err_str[PTL_MAX_ERRNO], "PTL_MAX_ERRNO"));
+
ptl_ni_init();
ptl_me_init();
ptl_eq_init();
space = nal->cb_malloc (nal, n * size);
if (space == NULL)
- return (PTL_NOSPACE);
+ return (PTL_NO_SPACE);
INIT_LIST_HEAD (&fl->fl_list);
fl->fl_objs = space;
(struct list_head *)nal->cb_malloc (nal, ni->ni_lh_hash_size
* sizeof (struct list_head));
if (ni->ni_lh_hash_table == NULL)
- return (PTL_NOSPACE);
+ return (PTL_NO_SPACE);
for (i = 0; i < ni->ni_lh_hash_size; i++)
INIT_LIST_HEAD (&ni->ni_lh_hash_table[i]);
ni->tbl.tbl = nal->cb_malloc(nal, sizeof(struct list_head) * ptl_size);
if (ni->tbl.tbl == NULL) {
- rc = PTL_NOSPACE;
+ rc = PTL_NO_SPACE;
goto out;
}
* otherwise caller may only lib_md_free() it.
*/
- if (!PtlHandleEqual (*eqh, PTL_EQ_NONE)) {
+ if (!PtlHandleIsEqual (*eqh, PTL_EQ_NONE)) {
eq = ptl_handle2eq(eqh, nal);
if (eq == NULL)
- return PTL_INV_EQ;
+ return PTL_EQ_INVALID;
}
/* Must check this _before_ allocation. Also, note that non-iov
* MDs must set md_niov to 0. */
- LASSERT((md->options & (PTL_MD_IOV | PTL_MD_KIOV)) == 0 ||
+ LASSERT((md->options & (PTL_MD_IOVEC | PTL_MD_KIOV)) == 0 ||
md->niov <= PTL_MD_MAX_IOV);
/* This implementation doesn't know how to create START events or
* disable END events. Best to LASSERT our caller is compliant so
* we find out quickly... */
- LASSERT (PtlHandleEqual (*eqh, PTL_EQ_NONE) ||
+ LASSERT (PtlHandleIsEqual (*eqh, PTL_EQ_NONE) ||
((md->options & PTL_MD_EVENT_START_DISABLE) != 0 &&
(md->options & PTL_MD_EVENT_END_DISABLE) == 0));
if ((md->options & PTL_MD_MAX_SIZE) != 0 && /* max size used */
(md->max_size < 0 || md->max_size > md->length)) // illegal max_size
- return PTL_INV_MD;
+ return PTL_MD_INVALID;
new->me = NULL;
new->start = md->start;
new->pending = 0;
new->md_flags = (unlink == PTL_UNLINK) ? PTL_MD_FLAG_AUTO_UNLINK : 0;
- if ((md->options & PTL_MD_IOV) != 0) {
+ if ((md->options & PTL_MD_IOVEC) != 0) {
int total_length = 0;
if ((md->options & PTL_MD_KIOV) != 0) /* Can't specify both */
- return PTL_INV_MD;
+ return PTL_MD_INVALID;
new->md_niov = md->niov;
}
} else if ((md->options & PTL_MD_KIOV) != 0) {
#ifndef __KERNEL__
- return PTL_INV_MD;
+ return PTL_MD_INVALID;
#else
int total_length = 0;
/* Trap attempt to use paged I/O if unsupported early. */
if (nal->cb_send_pages == NULL ||
nal->cb_recv_pages == NULL)
- return PTL_INV_MD;
+ return PTL_MD_INVALID;
new->md_niov = md->niov;
new->options = md->options;
new->user_ptr = md->user_ptr;
ptl_eq2handle(&new->eventq, md->eq);
- new->niov = ((md->options & (PTL_MD_IOV | PTL_MD_KIOV)) == 0) ? 0 : md->md_niov;
+ new->niov = ((md->options & (PTL_MD_IOVEC | PTL_MD_KIOV)) == 0) ? 0 : md->md_niov;
}
int do_PtlMDAttach(nal_cb_t * nal, void *private, void *v_args, void *v_ret)
lib_md_t *md;
unsigned long flags;
- if ((args->md_in.options & (PTL_MD_KIOV | PTL_MD_IOV)) != 0 &&
+ if ((args->md_in.options & (PTL_MD_KIOV | PTL_MD_IOVEC)) != 0 &&
args->md_in.niov > PTL_MD_MAX_IOV) /* too many fragments */
return (ret->rc = PTL_IOV_TOO_MANY);
md = lib_md_alloc(nal, &args->md_in);
if (md == NULL)
- return (ret->rc = PTL_NOSPACE);
+ return (ret->rc = PTL_NO_SPACE);
state_lock(nal, &flags);
me = ptl_handle2me(&args->me_in, nal);
if (me == NULL) {
- ret->rc = PTL_INV_ME;
+ ret->rc = PTL_ME_INVALID;
} else if (me->md != NULL) {
- ret->rc = PTL_INUSE;
+ ret->rc = PTL_ME_IN_USE;
} else {
ret->rc = lib_md_build(nal, md, private, &args->md_in,
&args->eq_in, args->unlink_in);
lib_md_t *md;
unsigned long flags;
- if ((args->md_in.options & (PTL_MD_KIOV | PTL_MD_IOV)) != 0 &&
+ if ((args->md_in.options & (PTL_MD_KIOV | PTL_MD_IOVEC)) != 0 &&
args->md_in.niov > PTL_MD_MAX_IOV) /* too many fragments */
return (ret->rc = PTL_IOV_TOO_MANY);
md = lib_md_alloc(nal, &args->md_in);
if (md == NULL)
- return (ret->rc = PTL_NOSPACE);
+ return (ret->rc = PTL_NO_SPACE);
state_lock(nal, &flags);
md = ptl_handle2md(&args->md_in, nal);
if (md == NULL) {
state_unlock(nal, &flags);
- return (ret->rc = PTL_INV_MD);
+ return (ret->rc = PTL_MD_INVALID);
}
/* If the MD is busy, lib_md_unlink just marks it for deletion, and
md = ptl_handle2md(&args->md_in, nal);
if (md == NULL) {
- ret->rc = PTL_INV_MD;
+ ret->rc = PTL_MD_INVALID;
goto out;
}
/* XXX fttb, the new MD must be the same type wrt fragmentation */
if (((new->options ^ md->options) &
- (PTL_MD_IOV | PTL_MD_KIOV)) != 0) {
- ret->rc = PTL_INV_MD;
+ (PTL_MD_IOVEC | PTL_MD_KIOV)) != 0) {
+ ret->rc = PTL_MD_INVALID;
goto out;
}
goto out;
}
- if (!PtlHandleEqual (args->testq_in, PTL_EQ_NONE)) {
+ if (!PtlHandleIsEqual (args->testq_in, PTL_EQ_NONE)) {
test_eq = ptl_handle2eq(&args->testq_in, nal);
if (test_eq == NULL) {
- ret->rc = PTL_INV_EQ;
+ ret->rc = PTL_EQ_INVALID;
goto out;
}
}
if (md->pending != 0) {
- ret->rc = PTL_NOUPDATE;
+ ret->rc = PTL_MD_NO_UPDATE;
goto out;
}
md->me = me;
} else {
- ret->rc = PTL_NOUPDATE;
+ ret->rc = PTL_MD_NO_UPDATE;
}
out:
/* save transaction info for later finalize and cleanup */
PORTAL_ALLOC(ktx, (sizeof(kscimacnal_tx_t)));
if (!ktx) {
- return PTL_NOSPACE;
+ return PTL_NO_SPACE;
}
ktx->ktx_nmapped = 0; /* Start with no mapped pages :) */
kscimacnal_txrelease, ktx);
if (!msg) {
PORTAL_FREE(ktx, (sizeof(kscimacnal_tx_t)));
- return PTL_NOSPACE;
+ return PTL_NO_SPACE;
}
mac_put_mblk(msg, sizeof(ptl_hdr_t));
lastblk=msg;
if(!newblk) {
mac_free_msg(msg);
PORTAL_FREE(ktx, (sizeof(kscimacnal_tx_t)));
- return PTL_NOSPACE;
+ return PTL_NO_SPACE;
}
mac_put_mblk(newblk, nob);
mac_link_mblk(lastblk, newblk);
unsigned long flags;
if (!ptl_init)
- RETURN(PTL_NOINIT);
+ RETURN(PTL_NO_INIT);
if (neq_in < 1)
- RETURN(PTL_INV_EQ);
+ RETURN(PTL_EQ_INVALID);
nal = ptl_hndl2nal(&eventqs_in[0]);
if (nal == NULL)
- RETURN(PTL_INV_EQ);
+ RETURN(PTL_EQ_INVALID);
nal->lock(nal, &flags);
if (i > 0 &&
ptl_hndl2nal(&eventqs_in[i]) != nal) {
nal->unlock(nal, &flags);
- RETURN (PTL_INV_EQ);
+ RETURN (PTL_EQ_INVALID);
}
/* size must be a power of 2 to handle a wrapped sequence # */
if (ptl_init)
return PTL_OK;
+ LASSERT(!strcmp(ptl_err_str[PTL_MAX_ERRNO], "PTL_MAX_ERRNO"));
+
ptl_ni_init();
ptl_me_init();
ptl_eq_init();
space = nal->cb_malloc (nal, n * size);
if (space == NULL)
- return (PTL_NOSPACE);
+ return (PTL_NO_SPACE);
INIT_LIST_HEAD (&fl->fl_list);
fl->fl_objs = space;
(struct list_head *)nal->cb_malloc (nal, ni->ni_lh_hash_size
* sizeof (struct list_head));
if (ni->ni_lh_hash_table == NULL)
- return (PTL_NOSPACE);
+ return (PTL_NO_SPACE);
for (i = 0; i < ni->ni_lh_hash_size; i++)
INIT_LIST_HEAD (&ni->ni_lh_hash_table[i]);
ni->tbl.tbl = nal->cb_malloc(nal, sizeof(struct list_head) * ptl_size);
if (ni->tbl.tbl == NULL) {
- rc = PTL_NOSPACE;
+ rc = PTL_NO_SPACE;
goto out;
}
* otherwise caller may only lib_md_free() it.
*/
- if (!PtlHandleEqual (*eqh, PTL_EQ_NONE)) {
+ if (!PtlHandleIsEqual (*eqh, PTL_EQ_NONE)) {
eq = ptl_handle2eq(eqh, nal);
if (eq == NULL)
- return PTL_INV_EQ;
+ return PTL_EQ_INVALID;
}
/* Must check this _before_ allocation. Also, note that non-iov
* MDs must set md_niov to 0. */
- LASSERT((md->options & (PTL_MD_IOV | PTL_MD_KIOV)) == 0 ||
+ LASSERT((md->options & (PTL_MD_IOVEC | PTL_MD_KIOV)) == 0 ||
md->niov <= PTL_MD_MAX_IOV);
/* This implementation doesn't know how to create START events or
* disable END events. Best to LASSERT our caller is compliant so
* we find out quickly... */
- LASSERT (PtlHandleEqual (*eqh, PTL_EQ_NONE) ||
+ LASSERT (PtlHandleIsEqual (*eqh, PTL_EQ_NONE) ||
((md->options & PTL_MD_EVENT_START_DISABLE) != 0 &&
(md->options & PTL_MD_EVENT_END_DISABLE) == 0));
if ((md->options & PTL_MD_MAX_SIZE) != 0 && /* max size used */
(md->max_size < 0 || md->max_size > md->length)) // illegal max_size
- return PTL_INV_MD;
+ return PTL_MD_INVALID;
new->me = NULL;
new->start = md->start;
new->pending = 0;
new->md_flags = (unlink == PTL_UNLINK) ? PTL_MD_FLAG_AUTO_UNLINK : 0;
- if ((md->options & PTL_MD_IOV) != 0) {
+ if ((md->options & PTL_MD_IOVEC) != 0) {
int total_length = 0;
if ((md->options & PTL_MD_KIOV) != 0) /* Can't specify both */
- return PTL_INV_MD;
+ return PTL_MD_INVALID;
new->md_niov = md->niov;
}
} else if ((md->options & PTL_MD_KIOV) != 0) {
#ifndef __KERNEL__
- return PTL_INV_MD;
+ return PTL_MD_INVALID;
#else
int total_length = 0;
/* Trap attempt to use paged I/O if unsupported early. */
if (nal->cb_send_pages == NULL ||
nal->cb_recv_pages == NULL)
- return PTL_INV_MD;
+ return PTL_MD_INVALID;
new->md_niov = md->niov;
new->options = md->options;
new->user_ptr = md->user_ptr;
ptl_eq2handle(&new->eventq, md->eq);
- new->niov = ((md->options & (PTL_MD_IOV | PTL_MD_KIOV)) == 0) ? 0 : md->md_niov;
+ new->niov = ((md->options & (PTL_MD_IOVEC | PTL_MD_KIOV)) == 0) ? 0 : md->md_niov;
}
int do_PtlMDAttach(nal_cb_t * nal, void *private, void *v_args, void *v_ret)
lib_md_t *md;
unsigned long flags;
- if ((args->md_in.options & (PTL_MD_KIOV | PTL_MD_IOV)) != 0 &&
+ if ((args->md_in.options & (PTL_MD_KIOV | PTL_MD_IOVEC)) != 0 &&
args->md_in.niov > PTL_MD_MAX_IOV) /* too many fragments */
return (ret->rc = PTL_IOV_TOO_MANY);
md = lib_md_alloc(nal, &args->md_in);
if (md == NULL)
- return (ret->rc = PTL_NOSPACE);
+ return (ret->rc = PTL_NO_SPACE);
state_lock(nal, &flags);
me = ptl_handle2me(&args->me_in, nal);
if (me == NULL) {
- ret->rc = PTL_INV_ME;
+ ret->rc = PTL_ME_INVALID;
} else if (me->md != NULL) {
- ret->rc = PTL_INUSE;
+ ret->rc = PTL_ME_IN_USE;
} else {
ret->rc = lib_md_build(nal, md, private, &args->md_in,
&args->eq_in, args->unlink_in);
lib_md_t *md;
unsigned long flags;
- if ((args->md_in.options & (PTL_MD_KIOV | PTL_MD_IOV)) != 0 &&
+ if ((args->md_in.options & (PTL_MD_KIOV | PTL_MD_IOVEC)) != 0 &&
args->md_in.niov > PTL_MD_MAX_IOV) /* too many fragments */
return (ret->rc = PTL_IOV_TOO_MANY);
md = lib_md_alloc(nal, &args->md_in);
if (md == NULL)
- return (ret->rc = PTL_NOSPACE);
+ return (ret->rc = PTL_NO_SPACE);
state_lock(nal, &flags);
md = ptl_handle2md(&args->md_in, nal);
if (md == NULL) {
state_unlock(nal, &flags);
- return (ret->rc = PTL_INV_MD);
+ return (ret->rc = PTL_MD_INVALID);
}
/* If the MD is busy, lib_md_unlink just marks it for deletion, and
md = ptl_handle2md(&args->md_in, nal);
if (md == NULL) {
- ret->rc = PTL_INV_MD;
+ ret->rc = PTL_MD_INVALID;
goto out;
}
/* XXX fttb, the new MD must be the same type wrt fragmentation */
if (((new->options ^ md->options) &
- (PTL_MD_IOV | PTL_MD_KIOV)) != 0) {
- ret->rc = PTL_INV_MD;
+ (PTL_MD_IOVEC | PTL_MD_KIOV)) != 0) {
+ ret->rc = PTL_MD_INVALID;
goto out;
}
goto out;
}
- if (!PtlHandleEqual (args->testq_in, PTL_EQ_NONE)) {
+ if (!PtlHandleIsEqual (args->testq_in, PTL_EQ_NONE)) {
test_eq = ptl_handle2eq(&args->testq_in, nal);
if (test_eq == NULL) {
- ret->rc = PTL_INV_EQ;
+ ret->rc = PTL_EQ_INVALID;
goto out;
}
}
if (md->pending != 0) {
- ret->rc = PTL_NOUPDATE;
+ ret->rc = PTL_MD_NO_UPDATE;
goto out;
}
md->me = me;
} else {
- ret->rc = PTL_NOUPDATE;
+ ret->rc = PTL_MD_NO_UPDATE;
}
out: