/*
* Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved.
* Use is subject to license terms.
+ *
+ * Copyright (c) 2012, Intel Corporation.
*/
/*
* This file is part of Lustre, http://www.lustre.org/
CTLTYPE_INT | CTLFLAG_RW , &ksocknal_tunables.ksnd_keepalive_intvl,
0, "keepalive_intvl");
-cfs_sysctl_table_t ksocknal_top_ctl_table [] = {
+struct ctl_table ksocknal_top_ctl_table [] = {
&sysctl__lnet_ksocknal,
&sysctl__lnet_ksocknal_timeout,
&sysctl__lnet_ksocknal_credits,
ksocknal_lib_tunables_init ()
{
ksocknal_tunables.ksnd_sysctl =
- cfs_register_sysctl_table (ksocknal_top_ctl_table, 0);
+ register_sysctl_table (ksocknal_top_ctl_table);
if (ksocknal_tunables.ksnd_sysctl == NULL)
return -ENOMEM;
void
ksocknal_lib_tunables_fini ()
{
- if (ksocknal_tunables.ksnd_sysctl != NULL)
- cfs_unregister_sysctl_table (ksocknal_tunables.ksnd_sysctl);
+ if (ksocknal_tunables.ksnd_sysctl != NULL)
+ unregister_sysctl_table (ksocknal_tunables.ksnd_sysctl);
}
#else
int
*/
#define KSOCKNAL_MAX_BUFFER (1152*1024)
-void
-ksocknal_lib_bind_irq (unsigned int irq)
-{
- return;
-}
-
-unsigned int
-ksocknal_lib_sock_irq (cfs_socket_t *sock)
-{
- return 0;
-}
-
int
ksocknal_lib_get_conn_addrs (ksock_conn_t *conn)
{
int i;
for (nob = i = 0; i < niov; i++) {
- scratchiov[i].iov_base = cfs_kmap(kiov[i].kiov_page) +
+ scratchiov[i].iov_base = kmap(kiov[i].kiov_page) +
kiov[i].kiov_offset;
nob += scratchiov[i].iov_len = kiov[i].kiov_len;
}
*/
rc = -sock_send(sock, &msg, MSG_DONTWAIT, &sndlen);
for (i = 0; i < niov; i++)
- cfs_kunmap(kiov[i].kiov_page);
+ kunmap(kiov[i].kiov_page);
if (rc == 0)
rc = sndlen;
return rc;
/* NB we can't trust socket ops to either consume our iovs
* or leave them alone. */
for (nob = i = 0; i < niov; i++) {
- scratchiov[i].iov_base = cfs_kmap(kiov[i].kiov_page) + \
+ scratchiov[i].iov_base = kmap(kiov[i].kiov_page) + \
kiov[i].kiov_offset;
nob += scratchiov[i].iov_len = kiov[i].kiov_len;
}
LASSERT (nob <= conn->ksnc_rx_nob_wanted);
rc = -sock_receive(C2B_SOCK(conn->ksnc_sock), &msg, MSG_DONTWAIT, &rcvlen);
for (i = 0; i < niov; i++)
- cfs_kunmap(kiov[i].kiov_page);
+ kunmap(kiov[i].kiov_page);
if (rc == 0)
rc = rcvlen;
return (rc);
CFS_DECL_NET_DATA;
for (nob = i = 0; i < niov; i++) {
- scratchiov[i].iov_base = cfs_kmap(kiov[i].kiov_page) +
+ scratchiov[i].iov_base = kmap(kiov[i].kiov_page) +
kiov[i].kiov_offset;
nob += scratchiov[i].iov_len = kiov[i].kiov_len;
}
CFS_NET_EX;
for (i = 0; i < niov; i++)
- cfs_kunmap(kiov[i].kiov_page);
+ kunmap(kiov[i].kiov_page);
if (rc != 0) {
if (suio.uio_resid != nob &&\
CFS_DECL_NET_DATA;
for (nob = i = 0; i < niov; i++) {
- scratchiov[i].iov_base = cfs_kmap(kiov[i].kiov_page) + kiov[i].kiov_offset;
+ scratchiov[i].iov_base = kmap(kiov[i].kiov_page) + kiov[i].kiov_offset;
nob += scratchiov[i].iov_len = kiov[i].kiov_len;
}
LASSERT (nob <= conn->ksnc_rx_nob_wanted);
CFS_NET_EX;
for (i = 0; i < niov; i++)
- cfs_kunmap(kiov[i].kiov_page);
+ kunmap(kiov[i].kiov_page);
if (rc){
if (ruio.uio_resid != nob && \