/* -*- mode: c; c-basic-offset: 8; indent-tabs-mode: nil; -*- * vim:expandtab:shiftwidth=8:tabstop=8: * * Copyright (c) 2004 Cluster File Systems, Inc. * * This file is part of Lustre, http://www.lustre.org. * * Lustre is free software; you can redistribute it and/or * modify it under the terms of version 2 of the GNU General Public * License as published by the Free Software Foundation. * * Lustre is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with Lustre; if not, write to the Free Software * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. */ #define DEBUG_SUBSYSTEM S_RPC #ifndef __KERNEL__ #include #include #include #endif #include #include #include #include #include #include "ptlrpc_internal.h" #ifdef __KERNEL__ #if !CRAY_PORTALS void ptlrpc_fill_bulk_md (ptl_md_t *md, struct ptlrpc_bulk_desc *desc) { LASSERT (desc->bd_iov_count <= PTLRPC_MAX_BRW_PAGES); LASSERT (!(md->options & (PTL_MD_IOVEC | PTL_MD_KIOV | PTL_MD_PHYS))); md->options |= PTL_MD_KIOV; md->start = &desc->bd_iov[0]; md->length = desc->bd_iov_count; } void ptlrpc_add_bulk_page(struct ptlrpc_bulk_desc *desc, struct page *page, int pageoffset, int len) { ptl_kiov_t *kiov = &desc->bd_iov[desc->bd_iov_count]; kiov->kiov_page = page; kiov->kiov_offset = pageoffset; kiov->kiov_len = len; desc->bd_iov_count++; } #else /* CRAY_PORTALS */ #ifdef PTL_MD_KIOV #error "Conflicting compilation directives" #endif void ptlrpc_fill_bulk_md (ptl_md_t *md, struct ptlrpc_bulk_desc *desc) { LASSERT (desc->bd_iov_count <= PTLRPC_MAX_BRW_PAGES); LASSERT (!(md->options & (PTL_MD_IOVEC | PTL_MD_PHYS))); md->options |= (PTL_MD_IOVEC | PTL_MD_PHYS); md->start = &desc->bd_iov[0]; md->length = desc->bd_iov_count; } void ptlrpc_add_bulk_page(struct ptlrpc_bulk_desc *desc, struct page *page, int pageoffset, int len) { ptl_md_iovec_t *iov = &desc->bd_iov[desc->bd_iov_count]; /* Should get a compiler warning if sizeof(physaddr) > sizeof(void *) */ iov->iov_base = (void *)(page_to_phys(page) + pageoffset); iov->iov_len = len; desc->bd_iov_count++; } #endif /* CRAY_PORTALS */ #else /* !__KERNEL__ */ void ptlrpc_fill_bulk_md(ptl_md_t *md, struct ptlrpc_bulk_desc *desc) { #if CRAY_PORTALS LASSERT (!(md->options & (PTL_MD_IOVEC | PTL_MD_PHYS))); LASSERT (desc->bd_iov_count == 1); #else LASSERT (!(md->options & (PTL_MD_IOVEC | PTL_MD_KIOV | PTL_MD_PHYS))); #endif if (desc->bd_iov_count == 1) { md->start = desc->bd_iov[0].iov_base; md->length = desc->bd_iov[0].iov_len; return; } md->options |= PTL_MD_IOVEC; md->start = &desc->bd_iov[0]; md->length = desc->bd_iov_count; } static int can_merge_iovs(ptl_md_iovec_t *existing, ptl_md_iovec_t *candidate) { if (existing->iov_base + existing->iov_len == candidate->iov_base) return 1; #if 0 /* Enable this section to provide earlier evidence of fragmented bulk */ CERROR("Can't merge iovs %p for %x, %p for %x\n", existing->iov_base, existing->iov_len, candidate->iov_base, candidate->iov_len); #endif return 0; } void ptlrpc_add_bulk_page(struct ptlrpc_bulk_desc *desc, struct page *page, int pageoffset, int len) { ptl_md_iovec_t *iov = &desc->bd_iov[desc->bd_iov_count]; iov->iov_base = page->addr + pageoffset; iov->iov_len = len; if (desc->bd_iov_count > 0 && can_merge_iovs(iov - 1, iov)) { (iov - 1)->iov_len += len; } else { desc->bd_iov_count++; } } #endif /* !__KERNEL__ */