X-Git-Url: https://git.whamcloud.com/?a=blobdiff_plain;f=lustre%2Fobdclass%2Fcl_page.c;h=d74c812be35e2f42cad4a720c58f6b259275ce38;hb=11c63d251e4d121bab0e97ccd69e9fd6d5ba38e4;hp=ff67ce0b3fd62dd750f4c243835236adf01bf23a;hpb=419016ac3e53e798453106ec04412a4843620916;p=fs%2Flustre-release.git diff --git a/lustre/obdclass/cl_page.c b/lustre/obdclass/cl_page.c index ff67ce0..d74c812 100644 --- a/lustre/obdclass/cl_page.c +++ b/lustre/obdclass/cl_page.c @@ -28,6 +28,8 @@ /* * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved. * Use is subject to license terms. + * + * Copyright (c) 2011, 2012, Whamcloud, Inc. */ /* * This file is part of Lustre, http://www.lustre.org/ @@ -189,7 +191,7 @@ EXPORT_SYMBOL(cl_page_lookup); */ int cl_page_gang_lookup(const struct lu_env *env, struct cl_object *obj, struct cl_io *io, pgoff_t start, pgoff_t end, - struct cl_page_list *queue) + cl_page_gang_cb_t cb, void *cbdata) { struct cl_object_header *hdr; struct cl_page *page; @@ -242,7 +244,7 @@ int cl_page_gang_lookup(const struct lu_env *env, struct cl_object *obj, */ cl_page_get_trust(page); lu_ref_add_atomic(&page->cp_reference, - "page_list", cfs_current()); + "gang_lookup", cfs_current()); pvec[j++] = page; } @@ -259,30 +261,16 @@ int cl_page_gang_lookup(const struct lu_env *env, struct cl_object *obj, for (i = 0; i < j; ++i) { page = pvec[i]; - if (res == CLP_GANG_OKAY) { - typeof(cl_page_own) *page_own; - - page_own = queue->pl_nr ? - cl_page_own_try : cl_page_own; - if (page_own(env, io, page) == 0) { - cl_page_list_add(queue, page); - } else if (page->cp_state != CPS_FREEING) { - /* cl_page_own() won't fail unless - * the page is being freed. */ - LASSERT(queue->pl_nr != 0); - res = CLP_GANG_AGAIN; - } - } + if (res == CLP_GANG_OKAY) + res = (*cb)(env, io, page, cbdata); lu_ref_del(&page->cp_reference, - "page_list", cfs_current()); + "gang_lookup", cfs_current()); cl_page_put(env, page); } if (nr < CLT_PVEC_SIZE || end_of_region) break; - /* if the number of pages is zero, this will mislead the caller - * that there is no page any more. */ - if (queue->pl_nr && cfs_need_resched()) + if (res == CLP_GANG_OKAY && cfs_need_resched()) res = CLP_GANG_RESCHED; if (res != CLP_GANG_OKAY) break; @@ -420,7 +408,7 @@ static struct cl_page *cl_page_find0(const struct lu_env *env, struct cl_site *site = cl_object_site(o); int err; - LINVRNT(type == CPT_CACHEABLE || type == CPT_TRANSIENT); + LASSERT(type == CPT_CACHEABLE || type == CPT_TRANSIENT); cfs_might_sleep(); ENTRY; @@ -557,7 +545,7 @@ static inline int cl_page_invariant(const struct cl_page *pg) * Either page is early in initialization (has neither child * nor parent yet), or it is in the object radix tree. */ - ergo(pg->cp_state < CPS_FREEING, + ergo(pg->cp_state < CPS_FREEING && pg->cp_type == CPT_CACHEABLE, (void *)radix_tree_lookup(&header->coh_tree, pg->cp_index) == pg || (child == NULL && parent == NULL)); @@ -634,7 +622,6 @@ static void cl_page_state_set0(const struct lu_env *env, static void cl_page_state_set(const struct lu_env *env, struct cl_page *page, enum cl_page_state state) { - PINVRNT(env, page, cl_page_invariant(page)); cl_page_state_set0(env, page, state); } @@ -1041,7 +1028,6 @@ EXPORT_SYMBOL(cl_page_own_try); void cl_page_assume(const struct lu_env *env, struct cl_io *io, struct cl_page *pg) { - PASSERT(env, pg, pg->cp_state < CPS_OWNED); PASSERT(env, pg, pg->cp_owner == NULL); PINVRNT(env, pg, cl_object_same(pg->cp_obj, io->ci_obj)); PINVRNT(env, pg, cl_page_invariant(pg)); @@ -1360,7 +1346,6 @@ void cl_page_completion(const struct lu_env *env, /* cl_page::cp_req already cleared by the caller (osc_completion()) */ PASSERT(env, pg, pg->cp_req == NULL); PASSERT(env, pg, pg->cp_state == cl_req_type_state(crt)); - PINVRNT(env, pg, cl_page_invariant(pg)); ENTRY; CL_PAGE_HEADER(D_TRACE, env, pg, "%d %d\n", crt, ioret); @@ -1374,6 +1359,7 @@ void cl_page_completion(const struct lu_env *env, (const struct lu_env *, const struct cl_page_slice *, int), ioret); if (anchor) { + LASSERT(cl_page_is_vmlocked(env, pg)); LASSERT(pg->cp_sync_io == anchor); pg->cp_sync_io = NULL; cl_sync_io_note(anchor, ioret); @@ -1475,6 +1461,16 @@ int cl_page_is_under_lock(const struct lu_env *env, struct cl_io *io, } EXPORT_SYMBOL(cl_page_is_under_lock); +static int page_prune_cb(const struct lu_env *env, struct cl_io *io, + struct cl_page *page, void *cbdata) +{ + cl_page_own(env, io, page); + cl_page_unmap(env, io, page); + cl_page_discard(env, io, page); + cl_page_disown(env, io, page); + return CLP_GANG_OKAY; +} + /** * Purges all cached pages belonging to the object \a obj. */ @@ -1483,12 +1479,10 @@ int cl_pages_prune(const struct lu_env *env, struct cl_object *clobj) struct cl_thread_info *info; struct cl_object *obj = cl_object_top(clobj); struct cl_io *io; - struct cl_page_list *plist; int result; ENTRY; info = cl_env_info(env); - plist = &info->clt_list; io = &info->clt_io; /* @@ -1503,18 +1497,8 @@ int cl_pages_prune(const struct lu_env *env, struct cl_object *clobj) } do { - cl_page_list_init(plist); result = cl_page_gang_lookup(env, obj, io, 0, CL_PAGE_EOF, - plist); - /* - * Since we're purging the pages of an object, we don't care - * the possible outcomes of the following functions. - */ - cl_page_list_unmap(env, io, plist); - cl_page_list_discard(env, io, plist); - cl_page_list_disown(env, io, plist); - cl_page_list_fini(env, plist); - + page_prune_cb, NULL); if (result == CLP_GANG_RESCHED) cfs_cond_resched(); } while (result != CLP_GANG_OKAY);