/* * GPL HEADER START * * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 only, * as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License version 2 for more details (a copy is included * in the LICENSE file that accompanied this code). * * You should have received a copy of the GNU General Public License * version 2 along with this program; If not, see * http://www.gnu.org/licenses/gpl-2.0.html * * GPL HEADER END */ /* * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved. * Use is subject to license terms. * * Copyright (c) 2011, 2017, Intel Corporation. */ /* * This file is part of Lustre, http://www.lustre.org/ * * lustre/mgc/mgc_request_server.c * * Author: Nathan Rutman */ #define DEBUG_SUBSYSTEM S_MGC #define D_MGC D_CONFIG /*|D_WARNING*/ #include #include #include #include #include #include #include #include #include #include #include #include #include "mgc_internal.h" static int mgc_local_llog_init(const struct lu_env *env, struct obd_device *obd, struct obd_device *disk) { struct llog_ctxt *ctxt; int rc; ENTRY; rc = llog_setup(env, obd, &obd->obd_olg, LLOG_CONFIG_ORIG_CTXT, disk, &llog_osd_ops); if (rc) RETURN(rc); ctxt = llog_get_context(obd, LLOG_CONFIG_ORIG_CTXT); LASSERT(ctxt); ctxt->loc_dir = obd->u.cli.cl_mgc_configs_dir; llog_ctxt_put(ctxt); RETURN(0); } static int mgc_local_llog_fini(const struct lu_env *env, struct obd_device *obd) { struct llog_ctxt *ctxt; ENTRY; ctxt = llog_get_context(obd, LLOG_CONFIG_ORIG_CTXT); llog_cleanup(env, ctxt); RETURN(0); } static int mgc_fs_setup(const struct lu_env *env, struct obd_device *obd, struct super_block *sb) { struct lustre_sb_info *lsi = s2lsi(sb); struct client_obd *cli = &obd->u.cli; struct lu_fid rfid, fid; struct dt_object *root, *dto; int rc = 0; ENTRY; LASSERT(lsi); LASSERT(lsi->lsi_dt_dev); /* The mgc fs exclusion mutex. Only one fs can be setup at a time. */ mutex_lock(&cli->cl_mgc_mutex); /* Setup the configs dir */ fid.f_seq = FID_SEQ_LOCAL_NAME; fid.f_oid = 1; fid.f_ver = 0; rc = local_oid_storage_init(env, lsi->lsi_dt_dev, &fid, &cli->cl_mgc_los); if (rc) GOTO(out_mutex, rc); rc = dt_root_get(env, lsi->lsi_dt_dev, &rfid); if (rc) GOTO(out_los, rc); root = dt_locate_at(env, lsi->lsi_dt_dev, &rfid, &cli->cl_mgc_los->los_dev->dd_lu_dev, NULL); if (unlikely(IS_ERR(root))) GOTO(out_los, rc = PTR_ERR(root)); dto = local_file_find_or_create(env, cli->cl_mgc_los, root, MOUNT_CONFIGS_DIR, S_IFDIR | 0755); dt_object_put_nocache(env, root); if (IS_ERR(dto)) GOTO(out_los, rc = PTR_ERR(dto)); cli->cl_mgc_configs_dir = dto; LASSERT(lsi->lsi_osd_exp->exp_obd->obd_lvfs_ctxt.dt); rc = mgc_local_llog_init(env, obd, lsi->lsi_osd_exp->exp_obd); if (rc) GOTO(out_llog, rc); /* We take an obd ref to insure that we can't get to mgc_cleanup * without calling mgc_fs_cleanup first. */ class_incref(obd, "mgc_fs", obd); /* We keep the cl_mgc_sem until mgc_fs_cleanup */ EXIT; out_llog: if (rc) { dt_object_put(env, cli->cl_mgc_configs_dir); cli->cl_mgc_configs_dir = NULL; } out_los: if (rc < 0) { local_oid_storage_fini(env, cli->cl_mgc_los); out_mutex: cli->cl_mgc_los = NULL; mutex_unlock(&cli->cl_mgc_mutex); } return rc; } static int mgc_fs_cleanup(const struct lu_env *env, struct obd_device *obd) { struct client_obd *cli = &obd->u.cli; ENTRY; LASSERT(cli->cl_mgc_los); mgc_local_llog_fini(env, obd); dt_object_put_nocache(env, cli->cl_mgc_configs_dir); cli->cl_mgc_configs_dir = NULL; local_oid_storage_fini(env, cli->cl_mgc_los); cli->cl_mgc_los = NULL; class_decref(obd, "mgc_fs", obd); mutex_unlock(&cli->cl_mgc_mutex); RETURN(0); } /* Send target_reg message to MGS */ static int mgc_target_register(struct obd_export *exp, struct mgs_target_info *mti) { size_t mti_len = offsetof(struct mgs_target_info, mti_nidlist); struct ptlrpc_request *req; struct mgs_target_info *req_mti, *rep_mti; int rc; ENTRY; req = ptlrpc_request_alloc(class_exp2cliimp(exp), &RQF_MGS_TARGET_REG); if (!req) RETURN(-ENOMEM); server_mti_print("mgc_target_register: req", mti); if (target_supports_large_nid(mti)) { mti_len += mti->mti_nid_count * LNET_NIDSTR_SIZE; req_capsule_set_size(&req->rq_pill, &RMF_MGS_TARGET_INFO, RCL_CLIENT, mti_len); req_capsule_set_size(&req->rq_pill, &RMF_MGS_TARGET_INFO, RCL_SERVER, mti_len); } rc = ptlrpc_request_pack(req, LUSTRE_MGS_VERSION, MGS_TARGET_REG); if (rc < 0) { ptlrpc_request_free(req); RETURN(rc); } req_mti = req_capsule_client_get(&req->rq_pill, &RMF_MGS_TARGET_INFO); if (!req_mti) { ptlrpc_req_finished(req); RETURN(-ENOMEM); } memcpy(req_mti, mti, mti_len); ptlrpc_request_set_replen(req); CDEBUG(D_MGC, "register %s\n", mti->mti_svname); /* Limit how long we will wait for the enqueue to complete */ req->rq_delay_limit = MGC_TARGET_REG_LIMIT; /* if the target needs to regenerate the config log in MGS, it's better * to use some longer limit to let MGC have time to change connection to * another MGS (or try again with the same MGS) for the target (server) * will fail and exit if the request expired due to delay limit. */ if (mti->mti_flags & (LDD_F_UPDATE | LDD_F_NEED_INDEX)) req->rq_delay_limit = MGC_TARGET_REG_LIMIT_MAX; rc = ptlrpc_queue_wait(req); if (ptlrpc_client_replied(req)) { rep_mti = req_capsule_server_get(&req->rq_pill, &RMF_MGS_TARGET_INFO); if (rep_mti) { mti_len = offsetof(struct mgs_target_info, mti_nidlist); if (target_supports_large_nid(mti)) mti_len += mti->mti_nid_count * LNET_NIDSTR_SIZE; memcpy(mti, rep_mti, mti_len); } } if (!rc) { CDEBUG(D_MGC, "register %s got index = %d\n", mti->mti_svname, mti->mti_stripe_index); server_mti_print("mgc_target_register: rep", mti); } ptlrpc_req_finished(req); RETURN(rc); } int mgc_set_info_async_server(const struct lu_env *env, struct obd_export *exp, u32 keylen, void *key, u32 vallen, void *val, struct ptlrpc_request_set *set) { int rc = -EINVAL; ENTRY; /* FIXME move this to mgc_process_config */ if (KEY_IS(KEY_REGISTER_TARGET)) { size_t mti_len = offsetof(struct mgs_target_info, mti_nidlist); struct mgs_target_info *mti = val; if (target_supports_large_nid(mti)) mti_len += mti->mti_nid_count * LNET_NIDSTR_SIZE; if (vallen != mti_len) RETURN(-EINVAL); CDEBUG(D_MGC, "register_target %s %#x\n", mti->mti_svname, mti->mti_flags); rc = mgc_target_register(exp, mti); RETURN(rc); } if (KEY_IS(KEY_SET_FS)) { struct super_block *sb = (struct super_block *)val; if (vallen != sizeof(struct super_block)) RETURN(-EINVAL); rc = mgc_fs_setup(env, exp->exp_obd, sb); RETURN(rc); } if (KEY_IS(KEY_CLEAR_FS)) { if (vallen != 0) RETURN(-EINVAL); rc = mgc_fs_cleanup(env, exp->exp_obd); RETURN(rc); } RETURN(rc); } int mgc_process_nodemap_log(struct obd_device *obd, struct config_llog_data *cld) { struct ptlrpc_connection *mgc_conn; struct ptlrpc_request *req = NULL; struct mgs_config_body *body; struct mgs_config_res *res; struct nodemap_config *new_config = NULL; struct lu_nodemap *recent_nodemap = NULL; struct ptlrpc_bulk_desc *desc; struct page **pages = NULL; u64 config_read_offset = 0; u8 nodemap_cur_pass = 0; int nrpages = 0; bool eof = true; bool mne_swab = false; int i; int ealen; int rc; ENTRY; mgc_conn = class_exp2cliimp(cld->cld_mgcexp)->imp_connection; /* don't need to get local config */ if (LNetIsPeerLocal(&mgc_conn->c_peer.nid)) GOTO(out, rc = 0); /* allocate buffer for bulk transfer. * if this is the first time for this mgs to read logs, * CONFIG_READ_NRPAGES_INIT will be used since it will read all logs * once; otherwise, it only reads increment of logs, this should be * small and CONFIG_READ_NRPAGES will be used. */ nrpages = CONFIG_READ_NRPAGES_INIT; OBD_ALLOC_PTR_ARRAY(pages, nrpages); if (!pages) GOTO(out, rc = -ENOMEM); for (i = 0; i < nrpages; i++) { pages[i] = alloc_page(GFP_KERNEL); if (!pages[i]) GOTO(out, rc = -ENOMEM); } again: if (config_read_offset == 0) { new_config = nodemap_config_alloc(); if (IS_ERR(new_config)) { rc = PTR_ERR(new_config); new_config = NULL; GOTO(out, rc); } } LASSERT(mutex_is_locked(&cld->cld_lock)); req = ptlrpc_request_alloc(class_exp2cliimp(cld->cld_mgcexp), &RQF_MGS_CONFIG_READ); if (!req) GOTO(out, rc = -ENOMEM); rc = ptlrpc_request_pack(req, LUSTRE_MGS_VERSION, MGS_CONFIG_READ); if (rc) GOTO(out, rc); /* pack request */ body = req_capsule_client_get(&req->rq_pill, &RMF_MGS_CONFIG_BODY); LASSERT(body); LASSERT(sizeof(body->mcb_name) > strlen(cld->cld_logname)); rc = strscpy(body->mcb_name, cld->cld_logname, sizeof(body->mcb_name)); if (rc < 0) GOTO(out, rc); body->mcb_offset = config_read_offset; body->mcb_type = cld->cld_type; body->mcb_bits = PAGE_SHIFT; body->mcb_units = nrpages; body->mcb_nm_cur_pass = nodemap_cur_pass; /* allocate bulk transfer descriptor */ desc = ptlrpc_prep_bulk_imp(req, nrpages, 1, PTLRPC_BULK_PUT_SINK, MGS_BULK_PORTAL, &ptlrpc_bulk_kiov_pin_ops); if (!desc) GOTO(out, rc = -ENOMEM); for (i = 0; i < nrpages; i++) desc->bd_frag_ops->add_kiov_frag(desc, pages[i], 0, PAGE_SIZE); ptlrpc_request_set_replen(req); rc = ptlrpc_queue_wait(req); if (rc) GOTO(out, rc); res = req_capsule_server_get(&req->rq_pill, &RMF_MGS_CONFIG_RES); if (!res) GOTO(out, rc = -EPROTO); config_read_offset = res->mcr_offset; eof = config_read_offset == II_END_OFF; nodemap_cur_pass = res->mcr_nm_cur_pass; ealen = sptlrpc_cli_unwrap_bulk_read(req, req->rq_bulk, 0); if (ealen < 0) GOTO(out, rc = ealen); if (ealen > nrpages << PAGE_SHIFT) GOTO(out, rc = -EINVAL); if (ealen == 0) { /* no logs transferred */ /* config changed since first read RPC */ if (config_read_offset == 0) { CDEBUG(D_INFO, "nodemap config changed in transit, retrying\n"); GOTO(out, rc = -EAGAIN); } if (!eof) rc = -EINVAL; GOTO(out, rc); } mne_swab = req_capsule_rep_need_swab(&req->rq_pill); /* When a nodemap config is received, we build a new nodemap config, * with new nodemap structs. We keep track of the most recently added * nodemap since the config is read ordered by nodemap_id, and so it * is likely that the next record will be related. Because access to * the nodemaps is single threaded until the nodemap_config is active, * we don't need to reference count with recent_nodemap, though * recent_nodemap should be set to NULL when the nodemap_config * is either destroyed or set active. */ for (i = 0; i < nrpages && ealen > 0; i++) { union lu_page *ptr; int rc2; ptr = kmap(pages[i]); rc2 = nodemap_process_idx_pages(new_config, ptr, &recent_nodemap); kunmap(pages[i]); if (rc2 < 0) { CWARN("%s: error processing %s log nodemap: rc = %d\n", obd->obd_name, cld->cld_logname, rc2); GOTO(out, rc = rc2); } ealen -= PAGE_SIZE; } out: if (req) { ptlrpc_req_finished(req); req = NULL; } if (rc == 0 && !eof) goto again; if (new_config) { /* recent_nodemap cannot be used after set_active/dealloc */ if (rc == 0) nodemap_config_set_active_mgc(new_config); else nodemap_config_dealloc(new_config); } if (pages) { for (i = 0; i < nrpages; i++) { if (!pages[i]) break; __free_page(pages[i]); } OBD_FREE_PTR_ARRAY(pages, nrpages); } return rc; } int mgc_process_config_server(struct obd_device *obd, size_t len, void *buf) { struct lustre_cfg *lcfg = buf; int rc = -ENOENT; ENTRY; switch (lcfg->lcfg_command) { case LCFG_LOV_ADD_OBD: { /* Overloading this cfg command: register a new target */ struct mgs_target_info *mti; if (LUSTRE_CFG_BUFLEN(lcfg, 1) != sizeof(struct mgs_target_info)) GOTO(out, rc = -EINVAL); mti = lustre_cfg_buf(lcfg, 1); CDEBUG(D_MGC, "add_target %s %#x\n", mti->mti_svname, mti->mti_flags); rc = mgc_target_register(obd->u.cli.cl_mgc_mgsexp, mti); break; } case LCFG_LOV_DEL_OBD: /* Unregister has no meaning at the moment. */ CERROR("lov_del_obd unimplemented\n"); rc = -EINVAL; break; } out: return rc; } int mgc_barrier_glimpse_ast(struct ldlm_lock *lock, void *data) { struct config_llog_data *cld = lock->l_ast_data; int rc; ENTRY; if (cld->cld_stopping) RETURN(-ENODEV); rc = barrier_handler(s2lsi(cld->cld_cfg.cfg_sb)->lsi_dt_dev, (struct ptlrpc_request *)data); RETURN(rc); } /* Copy a remote log locally */ static int mgc_llog_local_copy(const struct lu_env *env, struct obd_device *obd, struct llog_ctxt *rctxt, struct llog_ctxt *lctxt, char *logname) { char *temp_log; int rc; ENTRY; /* * NB: mgc_process_server_cfg_log() always needs valid local copy * and works only on it, so that defines the process: * - copy current local copy to temp_log using llog_backup() * - copy remote llog to logname using llog_backup() * - if failed then restore logname from backup */ OBD_ALLOC(temp_log, strlen(logname) + 2); if (!temp_log) RETURN(-ENOMEM); sprintf(temp_log, "%sT", logname); /* copy current local llog to temp_log */ rc = llog_backup(env, obd, lctxt, lctxt, logname, temp_log); if (rc < 0 && rc != -ENOENT) CWARN("%s: failed to backup local config %s: rc = %d\n", obd->obd_name, logname, rc); /* build new local llog */ rc = llog_backup(env, obd, rctxt, lctxt, logname, logname); if (rc == -ENOENT) { CWARN("%s: no remote llog for %s, check MGS config\n", obd->obd_name, logname); llog_erase(env, lctxt, NULL, logname); } else if (rc < 0) { /* error during backup, get local one back from the copy */ CWARN("%s: failed to copy new config %s: rc = %d\n", obd->obd_name, logname, rc); llog_backup(env, obd, lctxt, lctxt, temp_log, logname); } llog_erase(env, lctxt, NULL, temp_log); OBD_FREE(temp_log, strlen(logname) + 2); return rc; } int mgc_process_server_cfg_log(struct lu_env *env, struct llog_ctxt **ctxt, struct lustre_sb_info *lsi, struct obd_device *mgc, struct config_llog_data *cld, int local_only) { struct llog_ctxt *lctxt = llog_get_context(mgc, LLOG_CONFIG_ORIG_CTXT); struct client_obd *cli = &mgc->u.cli; int rc = 0; /* Copy the setup log locally if we can. Don't mess around if we're * running an MGS though (logs are already local). */ if (lctxt && lsi && IS_SERVER(lsi) && !IS_MGS(lsi) && cli->cl_mgc_configs_dir && lu2dt_dev(cli->cl_mgc_configs_dir->do_lu.lo_dev) == lsi->lsi_dt_dev) { if (!local_only && !lsi->lsi_dt_dev->dd_rdonly) { /* Only try to copy log if we have the lock. */ CDEBUG(D_INFO, "%s: copy local log %s\n", mgc->obd_name, cld->cld_logname); rc = mgc_llog_local_copy(env, mgc, *ctxt, lctxt, cld->cld_logname); if (!rc) lsi->lsi_flags &= ~LDD_F_NO_LOCAL_LOGS; } if (local_only || rc) { if (unlikely(lsi->lsi_flags & LDD_F_NO_LOCAL_LOGS) || rc) { CWARN("%s: local log %s are not valid and/or remote logs are not accessbile rc = %d\n", mgc->obd_name, cld->cld_logname, rc); GOTO(out_pop, rc = -EIO); } if (strcmp(cld->cld_logname, PARAMS_FILENAME) != 0 && llog_is_empty(env, lctxt, cld->cld_logname)) { LCONSOLE_ERROR_MSG(0x13a, "Failed to get MGS log %s and no local copy.\n", cld->cld_logname); GOTO(out_pop, rc = -ENOENT); } CDEBUG(D_MGC, "%s: Failed to get MGS log %s, using local copy for now, will try to update later.\n", mgc->obd_name, cld->cld_logname); rc = 0; } /* Now, whether we copied or not, start using the local llog. * If we failed to copy, we'll start using whatever the old * log has. */ llog_ctxt_put(*ctxt); *ctxt = lctxt; lctxt = NULL; } else if (local_only) { /* no local log at client side */ GOTO(out_pop, rc = -EIO); } rc = -EAGAIN; if (lsi && IS_SERVER(lsi) && !IS_MGS(lsi) && lsi->lsi_dt_dev->dd_rdonly) { struct llog_ctxt *rctxt; /* Under readonly mode, we may have no local copy or local * copy is incomplete, so try to use remote llog firstly. */ rctxt = llog_get_context(mgc, LLOG_CONFIG_REPL_CTXT); LASSERT(rctxt); rc = class_config_parse_llog(env, rctxt, cld->cld_logname, &cld->cld_cfg); llog_ctxt_put(rctxt); } out_pop: if (lctxt) __llog_ctxt_put(env, lctxt); return rc; }