X-Git-Url: https://git.whamcloud.com/?a=blobdiff_plain;ds=sidebyside;f=lustre%2Fptlrpc%2Fgss%2Fgss_svc_upcall.c;h=1c643ec2b46b5ac34c2d552c56d3efdc47b4f58d;hb=917655fc2938b90a9c246dd2d58408c42aa1658d;hp=a00b54172241491318a3f82cb938f1c302ba63eb;hpb=87383c55e74a219e72bcf861a2d2e81d978a927f;p=fs%2Flustre-release.git diff --git a/lustre/ptlrpc/gss/gss_svc_upcall.c b/lustre/ptlrpc/gss/gss_svc_upcall.c index a00b541..1c643ec 100644 --- a/lustre/ptlrpc/gss/gss_svc_upcall.c +++ b/lustre/ptlrpc/gss/gss_svc_upcall.c @@ -51,6 +51,7 @@ #include #include #include +#include #include #include #include @@ -68,12 +69,15 @@ #include "gss_err.h" #include "gss_internal.h" #include "gss_api.h" +#include "gss_crypto.h" #define GSS_SVC_UPCALL_TIMEOUT (20) -static spinlock_t __ctx_index_lock; +static DEFINE_SPINLOCK(__ctx_index_lock); static __u64 __ctx_index; +unsigned int krb5_allow_old_client_csum; + __u64 gss_get_next_ctx_index(void) { __u64 idx; @@ -159,6 +163,18 @@ static struct cache_detail rsi_cache; static struct rsi *rsi_update(struct rsi *new, struct rsi *old); static struct rsi *rsi_lookup(struct rsi *item); +#ifdef HAVE_CACHE_DETAIL_WRITERS +static inline int channel_users(struct cache_detail *cd) +{ + return atomic_read(&cd->writers); +} +#else +static inline int channel_users(struct cache_detail *cd) +{ + return atomic_read(&cd->readers); +} +#endif + static inline int rsi_hash(struct rsi *item) { return hash_mem((char *)item->in_handle.data, item->in_handle.len, @@ -720,85 +736,6 @@ static struct rsc *rsc_update(struct rsc *new, struct rsc *old) * rsc cache flush * ****************************************/ -typedef int rsc_entry_match(struct rsc *rscp, long data); - -static void rsc_flush(rsc_entry_match *match, long data) -{ -#ifdef HAVE_CACHE_HEAD_HLIST - struct cache_head *ch = NULL; - struct hlist_head *head; -#else - struct cache_head **ch; -#endif - struct rsc *rscp; - int n; - ENTRY; - - write_lock(&rsc_cache.hash_lock); - for (n = 0; n < RSC_HASHMAX; n++) { -#ifdef HAVE_CACHE_HEAD_HLIST - head = &rsc_cache.hash_table[n]; - hlist_for_each_entry(ch, head, cache_list) { - rscp = container_of(ch, struct rsc, h); -#else - for (ch = &rsc_cache.hash_table[n]; *ch;) { - rscp = container_of(*ch, struct rsc, h); -#endif - - if (!match(rscp, data)) { -#ifndef HAVE_CACHE_HEAD_HLIST - ch = &((*ch)->next); -#endif - continue; - } - - /* it seems simply set NEGATIVE doesn't work */ -#ifdef HAVE_CACHE_HEAD_HLIST - hlist_del_init(&ch->cache_list); -#else - *ch = (*ch)->next; - rscp->h.next = NULL; -#endif - cache_get(&rscp->h); - set_bit(CACHE_NEGATIVE, &rscp->h.flags); - COMPAT_RSC_PUT(&rscp->h, &rsc_cache); - rsc_cache.entries--; - } - } - write_unlock(&rsc_cache.hash_lock); - EXIT; -} - -static int match_uid(struct rsc *rscp, long uid) -{ - if ((int) uid == -1) - return 1; - return ((int) rscp->ctx.gsc_uid == (int) uid); -} - -static int match_target(struct rsc *rscp, long target) -{ - return (rscp->target == (struct obd_device *) target); -} - -static inline void rsc_flush_uid(int uid) -{ - if (uid == -1) - CWARN("flush all gss contexts...\n"); - - rsc_flush(match_uid, (long) uid); -} - -static inline void rsc_flush_target(struct obd_device *target) -{ - rsc_flush(match_target, (long) target); -} - -void gss_secsvc_flush(struct obd_device *target) -{ - rsc_flush_target(target); -} - static struct rsc *gss_svc_searchbyctx(rawobj_t *handle) { struct rsc rsci; @@ -1002,17 +939,17 @@ cache_check: if (first_check) { first_check = 0; - read_lock(&rsi_cache.hash_lock); + cache_read_lock(&rsi_cache); valid = test_bit(CACHE_VALID, &rsip->h.flags); if (valid == 0) set_current_state(TASK_INTERRUPTIBLE); - read_unlock(&rsi_cache.hash_lock); + cache_read_unlock(&rsi_cache); if (valid == 0) { - unsigned long jiffies; - jiffies = msecs_to_jiffies(MSEC_PER_SEC * - GSS_SVC_UPCALL_TIMEOUT); - schedule_timeout(jiffies); + unsigned long timeout; + + timeout = cfs_time_seconds(GSS_SVC_UPCALL_TIMEOUT); + schedule_timeout(timeout); } cache_get(&rsip->h); goto cache_check; @@ -1055,6 +992,20 @@ cache_check: grctx->src_ctx = &rsci->ctx; } + if (gw->gw_flags & LUSTRE_GSS_PACK_KCSUM) { + grctx->src_ctx->gsc_mechctx->hash_func = gss_digest_hash; + } else if (!strcmp(grctx->src_ctx->gsc_mechctx->mech_type->gm_name, + "krb5") && + !krb5_allow_old_client_csum) { + CWARN("%s: deny connection from '%s' due to missing 'krb_csum' feature, set 'sptlrpc.gss.krb5_allow_old_client_csum=1' to allow, but recommend client upgrade: rc = %d\n", + target->obd_name, libcfs_nid2str(req->rq_peer.nid), + -EPROTO); + GOTO(out, rc = SECSVC_DROP); + } else { + grctx->src_ctx->gsc_mechctx->hash_func = + gss_digest_hash_compat; + } + if (rawobj_dup(&rsci->ctx.gsc_rvs_hdl, rvs_hdl)) { CERROR("failed duplicate reverse handle\n"); GOTO(out, rc); @@ -1161,14 +1112,13 @@ int __init gss_init_svc_upcall(void) { int i, rc; - spin_lock_init(&__ctx_index_lock); /* * this helps reducing context index confliction. after server reboot, * conflicting request from clients might be filtered out by initial * sequence number checking, thus no chance to sent error notification * back to clients. */ - cfs_get_random_bytes(&__ctx_index, sizeof(__ctx_index)); + get_random_bytes(&__ctx_index, sizeof(__ctx_index)); rc = _cache_register_net(&rsi_cache, &init_net); if (rc != 0) @@ -1187,14 +1137,13 @@ int __init gss_init_svc_upcall(void) * Here we wait at minimum 1.5 seconds. */ for (i = 0; i < 6; i++) { - if (atomic_read(&rsi_cache.readers) > 0) + if (channel_users(&rsi_cache) > 0) break; set_current_state(TASK_UNINTERRUPTIBLE); - LASSERT(msecs_to_jiffies(MSEC_PER_SEC / 4) > 0); - schedule_timeout(msecs_to_jiffies(MSEC_PER_SEC / 4)); + schedule_timeout(cfs_time_seconds(1) / 4); } - if (atomic_read(&rsi_cache.readers) == 0) + if (channel_users(&rsi_cache) == 0) CWARN("Init channel is not opened by lsvcgssd, following " "request might be dropped until lsvcgssd is active\n");