X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=blobdiff_plain;f=lustre%2Fptlrpc%2Fgss%2Fgss_svc_upcall.c;h=eb5e49382c49ececc4cadf672423d01505a9b1f9;hp=e91a658f3ada0da6cfa48baaedf21e3422a3023b;hb=5d9d8eb41a7465d4810c82c08d97e70c3f3749c2;hpb=08aa217ce49aba1ded52e0f7adb8a607035123fd diff --git a/lustre/ptlrpc/gss/gss_svc_upcall.c b/lustre/ptlrpc/gss/gss_svc_upcall.c index e91a658..eb5e493 100644 --- a/lustre/ptlrpc/gss/gss_svc_upcall.c +++ b/lustre/ptlrpc/gss/gss_svc_upcall.c @@ -3,7 +3,7 @@ * * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved. * - * Copyright (c) 2011, 2012, Intel Corporation. + * Copyright (c) 2012, Intel Corporation. * * Author: Eric Mei */ @@ -121,13 +121,13 @@ static inline unsigned long hash_mem(char *buf, int length, int bits) #define RSI_HASHMASK (RSI_HASHMAX - 1) struct rsi { - struct cache_head h; - __u32 lustre_svc; - __u64 nid; - cfs_waitq_t waitq; - rawobj_t in_handle, in_token; - rawobj_t out_handle, out_token; - int major_status, minor_status; + struct cache_head h; + __u32 lustre_svc; + __u64 nid; + wait_queue_head_t waitq; + rawobj_t in_handle, in_token; + rawobj_t out_handle, out_token; + int major_status, minor_status; }; static struct cache_head *rsi_table[RSI_HASHMAX]; @@ -177,26 +177,24 @@ static void rsi_request(struct cache_detail *cd, (*bpp)[-1] = '\n'; } -#ifdef HAVE_CACHE_UPCALL static int rsi_upcall(struct cache_detail *cd, struct cache_head *h) { return sunrpc_cache_pipe_upcall(cd, h, rsi_request); } -#endif static inline void __rsi_init(struct rsi *new, struct rsi *item) { - new->out_handle = RAWOBJ_EMPTY; - new->out_token = RAWOBJ_EMPTY; + new->out_handle = RAWOBJ_EMPTY; + new->out_token = RAWOBJ_EMPTY; - new->in_handle = item->in_handle; - item->in_handle = RAWOBJ_EMPTY; - new->in_token = item->in_token; - item->in_token = RAWOBJ_EMPTY; + new->in_handle = item->in_handle; + item->in_handle = RAWOBJ_EMPTY; + new->in_token = item->in_token; + item->in_token = RAWOBJ_EMPTY; - new->lustre_svc = item->lustre_svc; - new->nid = item->nid; - cfs_waitq_init(&new->waitq); + new->lustre_svc = item->lustre_svc; + new->nid = item->nid; + init_waitqueue_head(&new->waitq); } static inline void __rsi_update(struct rsi *new, struct rsi *item) @@ -337,17 +335,17 @@ static int rsi_parse(struct cache_detail *cd, char *mesg, int mlen) rsip = rsi_update(&rsii, rsip); status = 0; out: - rsi_free(&rsii); - if (rsip) { - cfs_waitq_broadcast(&rsip->waitq); - cache_put(&rsip->h, &rsi_cache); - } else { - status = -ENOMEM; - } - - if (status) - CERROR("rsi parse error %d\n", status); - RETURN(status); + rsi_free(&rsii); + if (rsip) { + wake_up_all(&rsip->waitq); + cache_put(&rsip->h, &rsi_cache); + } else { + status = -ENOMEM; + } + + if (status) + CERROR("rsi parse error %d\n", status); + RETURN(status); } static struct cache_detail rsi_cache = { @@ -355,11 +353,7 @@ static struct cache_detail rsi_cache = { .hash_table = rsi_table, .name = "auth.sptlrpc.init", .cache_put = rsi_put, -#ifdef HAVE_CACHE_UPCALL .cache_upcall = rsi_upcall, -#else - .cache_request = rsi_request, -#endif .cache_parse = rsi_parse, .match = rsi_match, .init = rsi_init, @@ -850,22 +844,22 @@ static struct cache_deferred_req* cache_upcall_defer(struct cache_req *req) static struct cache_req cache_upcall_chandle = { cache_upcall_defer }; int gss_svc_upcall_handle_init(struct ptlrpc_request *req, - struct gss_svc_reqctx *grctx, - struct gss_wire_ctx *gw, - struct obd_device *target, - __u32 lustre_svc, - rawobj_t *rvs_hdl, - rawobj_t *in_token) + struct gss_svc_reqctx *grctx, + struct gss_wire_ctx *gw, + struct obd_device *target, + __u32 lustre_svc, + rawobj_t *rvs_hdl, + rawobj_t *in_token) { - struct ptlrpc_reply_state *rs; - struct rsc *rsci = NULL; - struct rsi *rsip = NULL, rsikey; - cfs_waitlink_t wait; - int replen = sizeof(struct ptlrpc_body); - struct gss_rep_header *rephdr; - int first_check = 1; - int rc = SECSVC_DROP; - ENTRY; + struct ptlrpc_reply_state *rs; + struct rsc *rsci = NULL; + struct rsi *rsip = NULL, rsikey; + wait_queue_t wait; + int replen = sizeof(struct ptlrpc_body); + struct gss_rep_header *rephdr; + int first_check = 1; + int rc = SECSVC_DROP; + ENTRY; memset(&rsikey, 0, sizeof(rsikey)); rsikey.lustre_svc = lustre_svc; @@ -894,18 +888,19 @@ int gss_svc_upcall_handle_init(struct ptlrpc_request *req, GOTO(out, rc); } - cache_get(&rsip->h); /* take an extra ref */ - cfs_waitq_init(&rsip->waitq); - cfs_waitlink_init(&wait); - cfs_waitq_add(&rsip->waitq, &wait); + cache_get(&rsip->h); /* take an extra ref */ + init_waitqueue_head(&rsip->waitq); + init_waitqueue_entry_current(&wait); + add_wait_queue(&rsip->waitq, &wait); cache_check: - /* Note each time cache_check() will drop a reference if return - * non-zero. We hold an extra reference on initial rsip, but must - * take care of following calls. */ - rc = cache_check(&rsi_cache, &rsip->h, &cache_upcall_chandle); - switch (rc) { - case -EAGAIN: { + /* Note each time cache_check() will drop a reference if return + * non-zero. We hold an extra reference on initial rsip, but must + * take care of following calls. */ + rc = cache_check(&rsi_cache, &rsip->h, &cache_upcall_chandle); + switch (rc) { + case -ETIMEDOUT: + case -EAGAIN: { int valid; if (first_check) { @@ -914,15 +909,15 @@ cache_check: read_lock(&rsi_cache.hash_lock); valid = test_bit(CACHE_VALID, &rsip->h.flags); if (valid == 0) - cfs_set_current_state(CFS_TASK_INTERRUPTIBLE); + set_current_state(TASK_INTERRUPTIBLE); read_unlock(&rsi_cache.hash_lock); - if (valid == 0) - cfs_schedule_timeout(GSS_SVC_UPCALL_TIMEOUT * - CFS_HZ); + if (valid == 0) + schedule_timeout(GSS_SVC_UPCALL_TIMEOUT * + HZ); - cache_get(&rsip->h); - goto cache_check; + cache_get(&rsip->h); + goto cache_check; } CWARN("waited %ds timeout, drop\n", GSS_SVC_UPCALL_TIMEOUT); break; @@ -933,17 +928,17 @@ cache_check: case 0: /* if not the first check, we have to release the extra * reference we just added on it. */ - if (!first_check) - cache_put(&rsip->h, &rsi_cache); - CDEBUG(D_SEC, "cache_check is good\n"); - break; - } + if (!first_check) + cache_put(&rsip->h, &rsi_cache); + CDEBUG(D_SEC, "cache_check is good\n"); + break; + } - cfs_waitq_del(&rsip->waitq, &wait); - cache_put(&rsip->h, &rsi_cache); + remove_wait_queue(&rsip->waitq, &wait); + cache_put(&rsip->h, &rsi_cache); - if (rc) - GOTO(out, rc = SECSVC_DROP); + if (rc) + GOTO(out, rc = SECSVC_DROP); rc = SECSVC_DROP; rsci = gss_svc_searchbyctx(&rsip->out_handle); @@ -1084,13 +1079,13 @@ int __init gss_init_svc_upcall(void) * upcall issued before the channel be opened thus nfsv4 cache code will * drop the request direclty, thus lead to unnecessary recovery time. * here we wait at miximum 1.5 seconds. */ - for (i = 0; i < 6; i++) { - if (atomic_read(&rsi_cache.readers) > 0) - break; - cfs_set_current_state(TASK_UNINTERRUPTIBLE); - LASSERT(CFS_HZ >= 4); - cfs_schedule_timeout(CFS_HZ / 4); - } + for (i = 0; i < 6; i++) { + if (atomic_read(&rsi_cache.readers) > 0) + break; + set_current_state(TASK_UNINTERRUPTIBLE); + LASSERT(HZ >= 4); + schedule_timeout(HZ / 4); + } if (atomic_read(&rsi_cache.readers) == 0) CWARN("Init channel is not opened by lsvcgssd, following "