/* * GPL HEADER START * * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 only, * as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License version 2 for more details (a copy is included * in the LICENSE file that accompanied this code). * * You should have received a copy of the GNU General Public License * version 2 along with this program; If not, see * http://www.sun.com/software/products/lustre/docs/GPLv2.pdf * * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, * CA 95054 USA or visit www.sun.com if you need additional information or * have any questions. * * GPL HEADER END */ /* * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved. * Use is subject to license terms. */ /* * This file is part of Lustre, http://www.lustre.org/ * Lustre is a trademark of Sun Microsystems, Inc. */ #ifndef _LUSTRE_SEC_H_ #define _LUSTRE_SEC_H_ /** \defgroup sptlrpc sptlrpc * * @{ */ /* * to avoid include */ struct obd_import; struct obd_export; struct ptlrpc_request; struct ptlrpc_reply_state; struct ptlrpc_bulk_desc; struct brw_page; /* Linux specific */ struct key; struct seq_file; /* * forward declaration */ struct ptlrpc_sec_policy; struct ptlrpc_sec_cops; struct ptlrpc_sec_sops; struct ptlrpc_sec; struct ptlrpc_svc_ctx; struct ptlrpc_cli_ctx; struct ptlrpc_ctx_ops; /** * \addtogroup flavor flavor * * RPC flavor is represented by a 32 bits integer. Currently the high 12 bits * are unused, must be set to 0 for future expansion. *
 * ------------------------------------------------------------------------
 * | 4b (bulk svc) | 4b (bulk type) | 4b (svc) | 4b (mech)  | 4b (policy) |
 * ------------------------------------------------------------------------
 * 
* * @{ */ /* * flavor constants */ enum sptlrpc_policy { SPTLRPC_POLICY_NULL = 0, SPTLRPC_POLICY_PLAIN = 1, SPTLRPC_POLICY_GSS = 2, SPTLRPC_POLICY_MAX, }; enum sptlrpc_mech_null { SPTLRPC_MECH_NULL = 0, SPTLRPC_MECH_NULL_MAX, }; enum sptlrpc_mech_plain { SPTLRPC_MECH_PLAIN = 0, SPTLRPC_MECH_PLAIN_MAX, }; enum sptlrpc_mech_gss { SPTLRPC_MECH_GSS_NULL = 0, SPTLRPC_MECH_GSS_KRB5 = 1, SPTLRPC_MECH_GSS_MAX, }; enum sptlrpc_service_type { SPTLRPC_SVC_NULL = 0, /**< no security */ SPTLRPC_SVC_AUTH = 1, /**< authentication only */ SPTLRPC_SVC_INTG = 2, /**< integrity */ SPTLRPC_SVC_PRIV = 3, /**< privacy */ SPTLRPC_SVC_MAX, }; enum sptlrpc_bulk_type { SPTLRPC_BULK_DEFAULT = 0, /**< follow rpc flavor */ SPTLRPC_BULK_HASH = 1, /**< hash integrity */ SPTLRPC_BULK_MAX, }; enum sptlrpc_bulk_service { SPTLRPC_BULK_SVC_NULL = 0, /**< no security */ SPTLRPC_BULK_SVC_AUTH = 1, /**< authentication only */ SPTLRPC_BULK_SVC_INTG = 2, /**< integrity */ SPTLRPC_BULK_SVC_PRIV = 3, /**< privacy */ SPTLRPC_BULK_SVC_MAX, }; /* * compose/extract macros */ #define FLVR_POLICY_OFFSET (0) #define FLVR_MECH_OFFSET (4) #define FLVR_SVC_OFFSET (8) #define FLVR_BULK_TYPE_OFFSET (12) #define FLVR_BULK_SVC_OFFSET (16) #define MAKE_FLVR(policy, mech, svc, btype, bsvc) \ (((__u32)(policy) << FLVR_POLICY_OFFSET) | \ ((__u32)(mech) << FLVR_MECH_OFFSET) | \ ((__u32)(svc) << FLVR_SVC_OFFSET) | \ ((__u32)(btype) << FLVR_BULK_TYPE_OFFSET) | \ ((__u32)(bsvc) << FLVR_BULK_SVC_OFFSET)) /* * extraction */ #define SPTLRPC_FLVR_POLICY(flavor) \ ((((__u32)(flavor)) >> FLVR_POLICY_OFFSET) & 0xF) #define SPTLRPC_FLVR_MECH(flavor) \ ((((__u32)(flavor)) >> FLVR_MECH_OFFSET) & 0xF) #define SPTLRPC_FLVR_SVC(flavor) \ ((((__u32)(flavor)) >> FLVR_SVC_OFFSET) & 0xF) #define SPTLRPC_FLVR_BULK_TYPE(flavor) \ ((((__u32)(flavor)) >> FLVR_BULK_TYPE_OFFSET) & 0xF) #define SPTLRPC_FLVR_BULK_SVC(flavor) \ ((((__u32)(flavor)) >> FLVR_BULK_SVC_OFFSET) & 0xF) #define SPTLRPC_FLVR_BASE(flavor) \ ((((__u32)(flavor)) >> FLVR_POLICY_OFFSET) & 0xFFF) #define SPTLRPC_FLVR_BASE_SUB(flavor) \ ((((__u32)(flavor)) >> FLVR_MECH_OFFSET) & 0xFF) /* * gss subflavors */ #define MAKE_BASE_SUBFLVR(mech, svc) \ ((__u32)(mech) | \ ((__u32)(svc) << (FLVR_SVC_OFFSET - FLVR_MECH_OFFSET))) #define SPTLRPC_SUBFLVR_KRB5N \ MAKE_BASE_SUBFLVR(SPTLRPC_MECH_GSS_KRB5, SPTLRPC_SVC_NULL) #define SPTLRPC_SUBFLVR_KRB5A \ MAKE_BASE_SUBFLVR(SPTLRPC_MECH_GSS_KRB5, SPTLRPC_SVC_AUTH) #define SPTLRPC_SUBFLVR_KRB5I \ MAKE_BASE_SUBFLVR(SPTLRPC_MECH_GSS_KRB5, SPTLRPC_SVC_INTG) #define SPTLRPC_SUBFLVR_KRB5P \ MAKE_BASE_SUBFLVR(SPTLRPC_MECH_GSS_KRB5, SPTLRPC_SVC_PRIV) /* * "end user" flavors */ #define SPTLRPC_FLVR_NULL \ MAKE_FLVR(SPTLRPC_POLICY_NULL, \ SPTLRPC_MECH_NULL, \ SPTLRPC_SVC_NULL, \ SPTLRPC_BULK_DEFAULT, \ SPTLRPC_BULK_SVC_NULL) #define SPTLRPC_FLVR_PLAIN \ MAKE_FLVR(SPTLRPC_POLICY_PLAIN, \ SPTLRPC_MECH_PLAIN, \ SPTLRPC_SVC_NULL, \ SPTLRPC_BULK_HASH, \ SPTLRPC_BULK_SVC_INTG) #define SPTLRPC_FLVR_KRB5N \ MAKE_FLVR(SPTLRPC_POLICY_GSS, \ SPTLRPC_MECH_GSS_KRB5, \ SPTLRPC_SVC_NULL, \ SPTLRPC_BULK_DEFAULT, \ SPTLRPC_BULK_SVC_NULL) #define SPTLRPC_FLVR_KRB5A \ MAKE_FLVR(SPTLRPC_POLICY_GSS, \ SPTLRPC_MECH_GSS_KRB5, \ SPTLRPC_SVC_AUTH, \ SPTLRPC_BULK_DEFAULT, \ SPTLRPC_BULK_SVC_NULL) #define SPTLRPC_FLVR_KRB5I \ MAKE_FLVR(SPTLRPC_POLICY_GSS, \ SPTLRPC_MECH_GSS_KRB5, \ SPTLRPC_SVC_INTG, \ SPTLRPC_BULK_DEFAULT, \ SPTLRPC_BULK_SVC_INTG) #define SPTLRPC_FLVR_KRB5P \ MAKE_FLVR(SPTLRPC_POLICY_GSS, \ SPTLRPC_MECH_GSS_KRB5, \ SPTLRPC_SVC_PRIV, \ SPTLRPC_BULK_DEFAULT, \ SPTLRPC_BULK_SVC_PRIV) #define SPTLRPC_FLVR_DEFAULT SPTLRPC_FLVR_NULL #define SPTLRPC_FLVR_INVALID ((__u32) 0xFFFFFFFF) #define SPTLRPC_FLVR_ANY ((__u32) 0xFFF00000) /** * extract the useful part from wire flavor */ #define WIRE_FLVR(wflvr) (((__u32) (wflvr)) & 0x000FFFFF) /** @} flavor */ static inline void flvr_set_svc(__u32 *flvr, __u32 svc) { LASSERT(svc < SPTLRPC_SVC_MAX); *flvr = MAKE_FLVR(SPTLRPC_FLVR_POLICY(*flvr), SPTLRPC_FLVR_MECH(*flvr), svc, SPTLRPC_FLVR_BULK_TYPE(*flvr), SPTLRPC_FLVR_BULK_SVC(*flvr)); } static inline void flvr_set_bulk_svc(__u32 *flvr, __u32 svc) { LASSERT(svc < SPTLRPC_BULK_SVC_MAX); *flvr = MAKE_FLVR(SPTLRPC_FLVR_POLICY(*flvr), SPTLRPC_FLVR_MECH(*flvr), SPTLRPC_FLVR_SVC(*flvr), SPTLRPC_FLVR_BULK_TYPE(*flvr), svc); } struct bulk_spec_hash { __u8 hash_alg; }; /** * Full description of flavors being used on a ptlrpc connection, include * both regular RPC and bulk transfer parts. */ struct sptlrpc_flavor { /** * wire flavor, should be renamed to sf_wire. */ __u32 sf_rpc; /** * general flags of PTLRPC_SEC_FL_* */ __u32 sf_flags; /** * rpc flavor specification */ union { /* nothing for now */ } u_rpc; /** * bulk flavor specification */ union { struct bulk_spec_hash hash; } u_bulk; }; /** * identify the RPC is generated from what part of Lustre. It's encoded into * RPC requests and to be checked by ptlrpc service. */ enum lustre_sec_part { LUSTRE_SP_CLI = 0, LUSTRE_SP_MDT, LUSTRE_SP_OST, LUSTRE_SP_MGC, LUSTRE_SP_MGS, LUSTRE_SP_ANY = 0xFF }; const char *sptlrpc_part2name(enum lustre_sec_part sp); enum lustre_sec_part sptlrpc_target_sec_part(struct obd_device *obd); /** * A rule specifies a flavor to be used by a ptlrpc connection between * two Lustre parts. */ struct sptlrpc_rule { __u32 sr_netid; /* LNET network ID */ __u8 sr_from; /* sec_part */ __u8 sr_to; /* sec_part */ __u16 sr_padding; struct sptlrpc_flavor sr_flvr; }; /** * A set of rules in memory. * * Rules are generated and stored on MGS, and propagated to MDT, OST, * and client when needed. */ struct sptlrpc_rule_set { int srs_nslot; int srs_nrule; struct sptlrpc_rule *srs_rules; }; int sptlrpc_parse_flavor(const char *str, struct sptlrpc_flavor *flvr); int sptlrpc_flavor_has_bulk(struct sptlrpc_flavor *flvr); static inline void sptlrpc_rule_set_init(struct sptlrpc_rule_set *set) { memset(set, 0, sizeof(*set)); } void sptlrpc_rule_set_free(struct sptlrpc_rule_set *set); int sptlrpc_rule_set_expand(struct sptlrpc_rule_set *set); int sptlrpc_rule_set_merge(struct sptlrpc_rule_set *set, struct sptlrpc_rule *rule); int sptlrpc_rule_set_choose(struct sptlrpc_rule_set *rset, enum lustre_sec_part from, enum lustre_sec_part to, lnet_nid_t nid, struct sptlrpc_flavor *sf); void sptlrpc_rule_set_dump(struct sptlrpc_rule_set *set); int sptlrpc_process_config(struct lustre_cfg *lcfg); void sptlrpc_conf_log_start(const char *logname); void sptlrpc_conf_log_stop(const char *logname); void sptlrpc_conf_log_update_begin(const char *logname); void sptlrpc_conf_log_update_end(const char *logname); void sptlrpc_conf_client_adapt(struct obd_device *obd); int sptlrpc_conf_target_get_rules(struct obd_device *obd, struct sptlrpc_rule_set *rset, int initial); void sptlrpc_target_choose_flavor(struct sptlrpc_rule_set *rset, enum lustre_sec_part from, lnet_nid_t nid, struct sptlrpc_flavor *flavor); /* The maximum length of security payload. 1024 is enough for Kerberos 5, * and should be enough for other future mechanisms but not sure. * Only used by pre-allocated request/reply pool. */ #define SPTLRPC_MAX_PAYLOAD (1024) struct vfs_cred { uint32_t vc_uid; uint32_t vc_gid; }; struct ptlrpc_ctx_ops { /** * To determine whether it's suitable to use the \a ctx for \a vcred. */ int (*match) (struct ptlrpc_cli_ctx *ctx, struct vfs_cred *vcred); /** * To bring the \a ctx uptodate. */ int (*refresh) (struct ptlrpc_cli_ctx *ctx); /** * Validate the \a ctx. */ int (*validate) (struct ptlrpc_cli_ctx *ctx); /** * Force the \a ctx to die. */ void (*die) (struct ptlrpc_cli_ctx *ctx, int grace); int (*display) (struct ptlrpc_cli_ctx *ctx, char *buf, int bufsize); /** * Sign the request message using \a ctx. * * \pre req->rq_reqmsg point to request message. * \pre req->rq_reqlen is the request message length. * \post req->rq_reqbuf point to request message with signature. * \post req->rq_reqdata_len is set to the final request message size. * * \see null_ctx_sign(), plain_ctx_sign(), gss_cli_ctx_sign(). */ int (*sign) (struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req); /** * Verify the reply message using \a ctx. * * \pre req->rq_repdata point to reply message with signature. * \pre req->rq_repdata_len is the total reply message length. * \post req->rq_repmsg point to reply message without signature. * \post req->rq_replen is the reply message length. * * \see null_ctx_verify(), plain_ctx_verify(), gss_cli_ctx_verify(). */ int (*verify) (struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req); /** * Encrypt the request message using \a ctx. * * \pre req->rq_reqmsg point to request message in clear text. * \pre req->rq_reqlen is the request message length. * \post req->rq_reqbuf point to request message. * \post req->rq_reqdata_len is set to the final request message size. * * \see gss_cli_ctx_seal(). */ int (*seal) (struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req); /** * Decrypt the reply message using \a ctx. * * \pre req->rq_repdata point to encrypted reply message. * \pre req->rq_repdata_len is the total cipher text length. * \post req->rq_repmsg point to reply message in clear text. * \post req->rq_replen is the reply message length in clear text. * * \see gss_cli_ctx_unseal(). */ int (*unseal) (struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req); /** * Wrap bulk request data. This is called before wrapping RPC * request message. * * \pre bulk buffer is descripted by desc->bd_iov and * desc->bd_iov_count. note for read it's just buffer, no data * need to be sent; for write it contains data in clear text. * \post when necessary, ptlrpc_bulk_sec_desc was properly prepared * (usually inside of RPC request message). * - encryption: cipher text bulk buffer is descripted by * desc->bd_enc_iov and desc->bd_iov_count (currently assume iov * count remains the same). * - otherwise: bulk buffer is still desc->bd_iov and * desc->bd_iov_count. * * \return 0: success. * \return -ev: error code. * * \see plain_cli_wrap_bulk(), gss_cli_ctx_wrap_bulk(). */ int (*wrap_bulk) (struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); /** * Unwrap bulk reply data. This is called after wrapping RPC * reply message. * * \pre bulk buffer is descripted by desc->bd_iov/desc->bd_enc_iov and * desc->bd_iov_count, according to wrap_bulk(). * \post final bulk data in clear text is placed in buffer described * by desc->bd_iov and desc->bd_iov_count. * \return +ve nob of actual bulk data in clear text. * \return -ve error code. * * \see plain_cli_unwrap_bulk(), gss_cli_ctx_unwrap_bulk(). */ int (*unwrap_bulk) (struct ptlrpc_cli_ctx *ctx, struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); }; #define PTLRPC_CTX_NEW_BIT (0) /* newly created */ #define PTLRPC_CTX_UPTODATE_BIT (1) /* uptodate */ #define PTLRPC_CTX_DEAD_BIT (2) /* mark expired gracefully */ #define PTLRPC_CTX_ERROR_BIT (3) /* fatal error (refresh, etc.) */ #define PTLRPC_CTX_CACHED_BIT (8) /* in ctx cache (hash etc.) */ #define PTLRPC_CTX_ETERNAL_BIT (9) /* always valid */ #define PTLRPC_CTX_NEW (1 << PTLRPC_CTX_NEW_BIT) #define PTLRPC_CTX_UPTODATE (1 << PTLRPC_CTX_UPTODATE_BIT) #define PTLRPC_CTX_DEAD (1 << PTLRPC_CTX_DEAD_BIT) #define PTLRPC_CTX_ERROR (1 << PTLRPC_CTX_ERROR_BIT) #define PTLRPC_CTX_CACHED (1 << PTLRPC_CTX_CACHED_BIT) #define PTLRPC_CTX_ETERNAL (1 << PTLRPC_CTX_ETERNAL_BIT) #define PTLRPC_CTX_STATUS_MASK (PTLRPC_CTX_NEW_BIT | \ PTLRPC_CTX_UPTODATE | \ PTLRPC_CTX_DEAD | \ PTLRPC_CTX_ERROR) struct ptlrpc_cli_ctx { cfs_hlist_node_t cc_cache; /* linked into ctx cache */ cfs_atomic_t cc_refcount; struct ptlrpc_sec *cc_sec; struct ptlrpc_ctx_ops *cc_ops; cfs_time_t cc_expire; /* in seconds */ unsigned int cc_early_expire:1; unsigned long cc_flags; struct vfs_cred cc_vcred; cfs_spinlock_t cc_lock; cfs_list_t cc_req_list; /* waiting reqs linked here */ cfs_list_t cc_gc_chain; /* linked to gc chain */ }; /** * client side policy operation vector. */ struct ptlrpc_sec_cops { /** * Given an \a imp, create and initialize a ptlrpc_sec structure. * \param ctx service context: * - regular import: \a ctx should be NULL; * - reverse import: \a ctx is obtained from incoming request. * \param flavor specify what flavor to use. * * When necessary, policy module is responsible for taking reference * on the import. * * \see null_create_sec(), plain_create_sec(), gss_sec_create_kr(). */ struct ptlrpc_sec * (*create_sec) (struct obd_import *imp, struct ptlrpc_svc_ctx *ctx, struct sptlrpc_flavor *flavor); /** * Destructor of ptlrpc_sec. When called, refcount has been dropped * to 0 and all contexts has been destroyed. * * \see null_destroy_sec(), plain_destroy_sec(), gss_sec_destroy_kr(). */ void (*destroy_sec) (struct ptlrpc_sec *sec); /** * Notify that this ptlrpc_sec is going to die. Optionally, policy * module is supposed to set sec->ps_dying and whatever necessary * actions. * * \see plain_kill_sec(), gss_sec_kill(). */ void (*kill_sec) (struct ptlrpc_sec *sec); /** * Given \a vcred, lookup and/or create its context. The policy module * is supposed to maintain its own context cache. * XXX currently \a create and \a remove_dead is always 1, perhaps * should be removed completely. * * \see null_lookup_ctx(), plain_lookup_ctx(), gss_sec_lookup_ctx_kr(). */ struct ptlrpc_cli_ctx * (*lookup_ctx) (struct ptlrpc_sec *sec, struct vfs_cred *vcred, int create, int remove_dead); /** * Called then the reference of \a ctx dropped to 0. The policy module * is supposed to destroy this context or whatever else according to * its cache maintainance mechamism. * * \param sync if zero, we shouldn't wait for the context being * destroyed completely. * * \see plain_release_ctx(), gss_sec_release_ctx_kr(). */ void (*release_ctx) (struct ptlrpc_sec *sec, struct ptlrpc_cli_ctx *ctx, int sync); /** * Flush the context cache. * * \param uid context of which user, -1 means all contexts. * \param grace if zero, the PTLRPC_CTX_UPTODATE_BIT of affected * contexts should be cleared immediately. * \param force if zero, only idle contexts will be flushed. * * \see plain_flush_ctx_cache(), gss_sec_flush_ctx_cache_kr(). */ int (*flush_ctx_cache) (struct ptlrpc_sec *sec, uid_t uid, int grace, int force); /** * Called periodically by garbage collector to remove dead contexts * from cache. * * \see gss_sec_gc_ctx_kr(). */ void (*gc_ctx) (struct ptlrpc_sec *sec); /** * Given an context \a ctx, install a corresponding reverse service * context on client side. * XXX currently it's only used by GSS module, maybe we should remove * this from general API. */ int (*install_rctx)(struct obd_import *imp, struct ptlrpc_sec *sec, struct ptlrpc_cli_ctx *ctx); /** * To allocate request buffer for \a req. * * \pre req->rq_reqmsg == NULL. * \pre req->rq_reqbuf == NULL, otherwise it must be pre-allocated, * we are not supposed to free it. * \post if success, req->rq_reqmsg point to a buffer with size * at least \a lustre_msg_size. * * \see null_alloc_reqbuf(), plain_alloc_reqbuf(), gss_alloc_reqbuf(). */ int (*alloc_reqbuf)(struct ptlrpc_sec *sec, struct ptlrpc_request *req, int lustre_msg_size); /** * To free request buffer for \a req. * * \pre req->rq_reqbuf != NULL. * * \see null_free_reqbuf(), plain_free_reqbuf(), gss_free_reqbuf(). */ void (*free_reqbuf) (struct ptlrpc_sec *sec, struct ptlrpc_request *req); /** * To allocate reply buffer for \a req. * * \pre req->rq_repbuf == NULL. * \post if success, req->rq_repbuf point to a buffer with size * req->rq_repbuf_len, the size should be large enough to receive * reply which be transformed from \a lustre_msg_size of clear text. * * \see null_alloc_repbuf(), plain_alloc_repbuf(), gss_alloc_repbuf(). */ int (*alloc_repbuf)(struct ptlrpc_sec *sec, struct ptlrpc_request *req, int lustre_msg_size); /** * To free reply buffer for \a req. * * \pre req->rq_repbuf != NULL. * \post req->rq_repbuf == NULL. * \post req->rq_repbuf_len == 0. * * \see null_free_repbuf(), plain_free_repbuf(), gss_free_repbuf(). */ void (*free_repbuf) (struct ptlrpc_sec *sec, struct ptlrpc_request *req); /** * To expand the request buffer of \a req, thus the \a segment in * the request message pointed by req->rq_reqmsg can accommodate * at least \a newsize of data. * * \pre req->rq_reqmsg->lm_buflens[segment] < newsize. * * \see null_enlarge_reqbuf(), plain_enlarge_reqbuf(), * gss_enlarge_reqbuf(). */ int (*enlarge_reqbuf) (struct ptlrpc_sec *sec, struct ptlrpc_request *req, int segment, int newsize); /* * misc */ int (*display) (struct ptlrpc_sec *sec, struct seq_file *seq); }; /** * server side policy operation vector. */ struct ptlrpc_sec_sops { /** * verify an incoming request. * * \pre request message is pointed by req->rq_reqbuf, size is * req->rq_reqdata_len; and the message has been unpacked to * host byte order. * * \retval SECSVC_OK success, req->rq_reqmsg point to request message * in clear text, size is req->rq_reqlen; req->rq_svc_ctx is set; * req->rq_sp_from is decoded from request. * \retval SECSVC_COMPLETE success, the request has been fully * processed, and reply message has been prepared; req->rq_sp_from is * decoded from request. * \retval SECSVC_DROP failed, this request should be dropped. * * \see null_accept(), plain_accept(), gss_svc_accept_kr(). */ int (*accept) (struct ptlrpc_request *req); /** * Perform security transformation upon reply message. * * \pre reply message is pointed by req->rq_reply_state->rs_msg, size * is req->rq_replen. * \post req->rs_repdata_len is the final message size. * \post req->rq_reply_off is set. * * \see null_authorize(), plain_authorize(), gss_svc_authorize(). */ int (*authorize) (struct ptlrpc_request *req); /** * Invalidate server context \a ctx. * * \see gss_svc_invalidate_ctx(). */ void (*invalidate_ctx) (struct ptlrpc_svc_ctx *ctx); /** * Allocate a ptlrpc_reply_state. * * \param msgsize size of the reply message in clear text. * \pre if req->rq_reply_state != NULL, then it's pre-allocated, we * should simply use it; otherwise we'll responsible for allocating * a new one. * \post req->rq_reply_state != NULL; * \post req->rq_reply_state->rs_msg != NULL; * * \see null_alloc_rs(), plain_alloc_rs(), gss_svc_alloc_rs(). */ int (*alloc_rs) (struct ptlrpc_request *req, int msgsize); /** * Free a ptlrpc_reply_state. */ void (*free_rs) (struct ptlrpc_reply_state *rs); /** * Release the server context \a ctx. * * \see gss_svc_free_ctx(). */ void (*free_ctx) (struct ptlrpc_svc_ctx *ctx); /** * Install a reverse context based on the server context \a ctx. * * \see gss_svc_install_rctx_kr(). */ int (*install_rctx)(struct obd_import *imp, struct ptlrpc_svc_ctx *ctx); /** * Prepare buffer for incoming bulk write. * * \pre desc->bd_iov and desc->bd_iov_count describes the buffer * intended to receive the write. * * \see gss_svc_prep_bulk(). */ int (*prep_bulk) (struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); /** * Unwrap the bulk write data. * * \see plain_svc_unwrap_bulk(), gss_svc_unwrap_bulk(). */ int (*unwrap_bulk) (struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); /** * Wrap the bulk read data. * * \see plain_svc_wrap_bulk(), gss_svc_wrap_bulk(). */ int (*wrap_bulk) (struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); }; struct ptlrpc_sec_policy { cfs_module_t *sp_owner; char *sp_name; __u16 sp_policy; /* policy number */ struct ptlrpc_sec_cops *sp_cops; /* client ops */ struct ptlrpc_sec_sops *sp_sops; /* server ops */ }; #define PTLRPC_SEC_FL_REVERSE 0x0001 /* reverse sec */ #define PTLRPC_SEC_FL_ROOTONLY 0x0002 /* treat everyone as root */ #define PTLRPC_SEC_FL_UDESC 0x0004 /* ship udesc */ #define PTLRPC_SEC_FL_BULK 0x0008 /* intensive bulk i/o expected */ #define PTLRPC_SEC_FL_PAG 0x0010 /* PAG mode */ /** * The ptlrpc_sec represents the client side ptlrpc security facilities, * each obd_import (both regular and reverse import) must associate with * a ptlrpc_sec. * * \see sptlrpc_import_sec_adapt(). */ struct ptlrpc_sec { struct ptlrpc_sec_policy *ps_policy; cfs_atomic_t ps_refcount; /** statistic only */ cfs_atomic_t ps_nctx; /** unique identifier */ int ps_id; struct sptlrpc_flavor ps_flvr; enum lustre_sec_part ps_part; /** after set, no more new context will be created */ unsigned int ps_dying:1; /** owning import */ struct obd_import *ps_import; cfs_spinlock_t ps_lock; /* * garbage collection */ cfs_list_t ps_gc_list; cfs_time_t ps_gc_interval; /* in seconds */ cfs_time_t ps_gc_next; /* in seconds */ }; static inline int sec_is_reverse(struct ptlrpc_sec *sec) { return (sec->ps_flvr.sf_flags & PTLRPC_SEC_FL_REVERSE); } static inline int sec_is_rootonly(struct ptlrpc_sec *sec) { return (sec->ps_flvr.sf_flags & PTLRPC_SEC_FL_ROOTONLY); } struct ptlrpc_svc_ctx { cfs_atomic_t sc_refcount; struct ptlrpc_sec_policy *sc_policy; }; /* * user identity descriptor */ #define LUSTRE_MAX_GROUPS (128) struct ptlrpc_user_desc { __u32 pud_uid; __u32 pud_gid; __u32 pud_fsuid; __u32 pud_fsgid; __u32 pud_cap; __u32 pud_ngroups; __u32 pud_groups[0]; }; /* * bulk flavors */ enum sptlrpc_bulk_hash_alg { BULK_HASH_ALG_NULL = 0, BULK_HASH_ALG_ADLER32, BULK_HASH_ALG_CRC32, BULK_HASH_ALG_MD5, BULK_HASH_ALG_SHA1, BULK_HASH_ALG_SHA256, BULK_HASH_ALG_SHA384, BULK_HASH_ALG_SHA512, BULK_HASH_ALG_MAX }; const char * sptlrpc_get_hash_name(__u8 hash_alg); __u8 sptlrpc_get_hash_alg(const char *algname); enum { BSD_FL_ERR = 1, }; struct ptlrpc_bulk_sec_desc { __u8 bsd_version; /* 0 */ __u8 bsd_type; /* SPTLRPC_BULK_XXX */ __u8 bsd_svc; /* SPTLRPC_BULK_SVC_XXXX */ __u8 bsd_flags; /* flags */ __u32 bsd_nob; /* nob of bulk data */ __u8 bsd_data[0]; /* policy-specific token */ }; /* * lprocfs */ struct proc_dir_entry; extern struct proc_dir_entry *sptlrpc_proc_root; /* * round size up to next power of 2, for slab allocation. * @size must be sane (can't overflow after round up) */ static inline int size_roundup_power2(int size) { size--; size |= size >> 1; size |= size >> 2; size |= size >> 4; size |= size >> 8; size |= size >> 16; size++; return size; } /* * internal support libraries */ void _sptlrpc_enlarge_msg_inplace(struct lustre_msg *msg, int segment, int newsize); /* * security policies */ int sptlrpc_register_policy(struct ptlrpc_sec_policy *policy); int sptlrpc_unregister_policy(struct ptlrpc_sec_policy *policy); __u32 sptlrpc_name2flavor_base(const char *name); const char *sptlrpc_flavor2name_base(__u32 flvr); char *sptlrpc_flavor2name_bulk(struct sptlrpc_flavor *sf, char *buf, int bufsize); char *sptlrpc_flavor2name(struct sptlrpc_flavor *sf, char *buf, int bufsize); char *sptlrpc_secflags2str(__u32 flags, char *buf, int bufsize); static inline struct ptlrpc_sec_policy *sptlrpc_policy_get(struct ptlrpc_sec_policy *policy) { __cfs_module_get(policy->sp_owner); return policy; } static inline void sptlrpc_policy_put(struct ptlrpc_sec_policy *policy) { cfs_module_put(policy->sp_owner); } /* * client credential */ static inline unsigned long cli_ctx_status(struct ptlrpc_cli_ctx *ctx) { return (ctx->cc_flags & PTLRPC_CTX_STATUS_MASK); } static inline int cli_ctx_is_ready(struct ptlrpc_cli_ctx *ctx) { return (cli_ctx_status(ctx) == PTLRPC_CTX_UPTODATE); } static inline int cli_ctx_is_refreshed(struct ptlrpc_cli_ctx *ctx) { return (cli_ctx_status(ctx) != 0); } static inline int cli_ctx_is_uptodate(struct ptlrpc_cli_ctx *ctx) { return ((ctx->cc_flags & PTLRPC_CTX_UPTODATE) != 0); } static inline int cli_ctx_is_error(struct ptlrpc_cli_ctx *ctx) { return ((ctx->cc_flags & PTLRPC_CTX_ERROR) != 0); } static inline int cli_ctx_is_dead(struct ptlrpc_cli_ctx *ctx) { return ((ctx->cc_flags & (PTLRPC_CTX_DEAD | PTLRPC_CTX_ERROR)) != 0); } static inline int cli_ctx_is_eternal(struct ptlrpc_cli_ctx *ctx) { return ((ctx->cc_flags & PTLRPC_CTX_ETERNAL) != 0); } /* * sec get/put */ struct ptlrpc_sec *sptlrpc_sec_get(struct ptlrpc_sec *sec); void sptlrpc_sec_put(struct ptlrpc_sec *sec); /* * internal apis which only used by policy impelentation */ int sptlrpc_get_next_secid(void); void sptlrpc_sec_destroy(struct ptlrpc_sec *sec); /* * exported client context api */ struct ptlrpc_cli_ctx *sptlrpc_cli_ctx_get(struct ptlrpc_cli_ctx *ctx); void sptlrpc_cli_ctx_put(struct ptlrpc_cli_ctx *ctx, int sync); void sptlrpc_cli_ctx_expire(struct ptlrpc_cli_ctx *ctx); void sptlrpc_cli_ctx_wakeup(struct ptlrpc_cli_ctx *ctx); int sptlrpc_cli_ctx_display(struct ptlrpc_cli_ctx *ctx, char *buf, int bufsize); /* * exported client context wrap/buffers */ int sptlrpc_cli_wrap_request(struct ptlrpc_request *req); int sptlrpc_cli_unwrap_reply(struct ptlrpc_request *req); int sptlrpc_cli_alloc_reqbuf(struct ptlrpc_request *req, int msgsize); void sptlrpc_cli_free_reqbuf(struct ptlrpc_request *req); int sptlrpc_cli_alloc_repbuf(struct ptlrpc_request *req, int msgsize); void sptlrpc_cli_free_repbuf(struct ptlrpc_request *req); int sptlrpc_cli_enlarge_reqbuf(struct ptlrpc_request *req, int segment, int newsize); int sptlrpc_cli_unwrap_early_reply(struct ptlrpc_request *req, struct ptlrpc_request **req_ret); void sptlrpc_cli_finish_early_reply(struct ptlrpc_request *early_req); void sptlrpc_request_out_callback(struct ptlrpc_request *req); /* * exported higher interface of import & request */ int sptlrpc_import_sec_adapt(struct obd_import *imp, struct ptlrpc_svc_ctx *ctx, struct sptlrpc_flavor *flvr); struct ptlrpc_sec *sptlrpc_import_sec_ref(struct obd_import *imp); void sptlrpc_import_sec_put(struct obd_import *imp); int sptlrpc_import_check_ctx(struct obd_import *imp); void sptlrpc_import_flush_root_ctx(struct obd_import *imp); void sptlrpc_import_flush_my_ctx(struct obd_import *imp); void sptlrpc_import_flush_all_ctx(struct obd_import *imp); int sptlrpc_req_get_ctx(struct ptlrpc_request *req); void sptlrpc_req_put_ctx(struct ptlrpc_request *req, int sync); int sptlrpc_req_refresh_ctx(struct ptlrpc_request *req, long timeout); int sptlrpc_req_replace_dead_ctx(struct ptlrpc_request *req); void sptlrpc_req_set_flavor(struct ptlrpc_request *req, int opcode); int sptlrpc_parse_rule(char *param, struct sptlrpc_rule *rule); /* gc */ void sptlrpc_gc_add_sec(struct ptlrpc_sec *sec); void sptlrpc_gc_del_sec(struct ptlrpc_sec *sec); void sptlrpc_gc_add_ctx(struct ptlrpc_cli_ctx *ctx); /* misc */ const char * sec2target_str(struct ptlrpc_sec *sec); int sptlrpc_lprocfs_cliobd_attach(struct obd_device *dev); /* * server side */ enum secsvc_accept_res { SECSVC_OK = 0, SECSVC_COMPLETE, SECSVC_DROP, }; int sptlrpc_svc_unwrap_request(struct ptlrpc_request *req); int sptlrpc_svc_alloc_rs(struct ptlrpc_request *req, int msglen); int sptlrpc_svc_wrap_reply(struct ptlrpc_request *req); void sptlrpc_svc_free_rs(struct ptlrpc_reply_state *rs); void sptlrpc_svc_ctx_addref(struct ptlrpc_request *req); void sptlrpc_svc_ctx_decref(struct ptlrpc_request *req); void sptlrpc_svc_ctx_invalidate(struct ptlrpc_request *req); int sptlrpc_target_export_check(struct obd_export *exp, struct ptlrpc_request *req); void sptlrpc_target_update_exp_flavor(struct obd_device *obd, struct sptlrpc_rule_set *rset); /* * reverse context */ int sptlrpc_svc_install_rvs_ctx(struct obd_import *imp, struct ptlrpc_svc_ctx *ctx); int sptlrpc_cli_install_rvs_ctx(struct obd_import *imp, struct ptlrpc_cli_ctx *ctx); /* bulk security api */ int sptlrpc_enc_pool_add_user(void); int sptlrpc_enc_pool_del_user(void); int sptlrpc_enc_pool_get_pages(struct ptlrpc_bulk_desc *desc); void sptlrpc_enc_pool_put_pages(struct ptlrpc_bulk_desc *desc); int sptlrpc_cli_wrap_bulk(struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); int sptlrpc_cli_unwrap_bulk_read(struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc, int nob); int sptlrpc_cli_unwrap_bulk_write(struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); #ifdef HAVE_SERVER_SUPPORT int sptlrpc_svc_prep_bulk(struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); int sptlrpc_svc_wrap_bulk(struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); int sptlrpc_svc_unwrap_bulk(struct ptlrpc_request *req, struct ptlrpc_bulk_desc *desc); #endif /* bulk helpers (internal use only by policies) */ int sptlrpc_get_bulk_checksum(struct ptlrpc_bulk_desc *desc, __u8 alg, void *buf, int buflen); int bulk_sec_desc_unpack(struct lustre_msg *msg, int offset, int swabbed); /* user descriptor helpers */ static inline int sptlrpc_user_desc_size(int ngroups) { return sizeof(struct ptlrpc_user_desc) + ngroups * sizeof(__u32); } int sptlrpc_current_user_desc_size(void); int sptlrpc_pack_user_desc(struct lustre_msg *msg, int offset); int sptlrpc_unpack_user_desc(struct lustre_msg *req, int offset, int swabbed); #define CFS_CAP_CHOWN_MASK (1 << CFS_CAP_CHOWN) #define CFS_CAP_SYS_RESOURCE_MASK (1 << CFS_CAP_SYS_RESOURCE) enum { LUSTRE_SEC_NONE = 0, LUSTRE_SEC_REMOTE = 1, LUSTRE_SEC_SPECIFY = 2, LUSTRE_SEC_ALL = 3 }; /** @} sptlrpc */ #endif /* _LUSTRE_SEC_H_ */