/* * GPL HEADER START * * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 only, * as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License version 2 for more details (a copy is included * in the LICENSE file that accompanied this code). * * You should have received a copy of the GNU General Public License * version 2 along with this program; If not, see * http://www.sun.com/software/products/lustre/docs/GPLv2.pdf * * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, * CA 95054 USA or visit www.sun.com if you need additional information or * have any questions. * * GPL HEADER END */ /* * Copyright (c) 2004, 2010, Oracle and/or its affiliates. All rights reserved. * Use is subject to license terms. * * Copyright (c) 2011, 2015, Intel Corporation. */ /* * This file is part of Lustre, http://www.lustre.org/ * Lustre is a trademark of Sun Microsystems, Inc. */ #define DEBUG_SUBSYSTEM S_LMV #include #include #include #include #ifdef HAVE_UIDGID_HEADER # include #endif #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "lmv_internal.h" static void lmv_activate_target(struct lmv_obd *lmv, struct lmv_tgt_desc *tgt, int activate) { if (tgt->ltd_active == activate) return; tgt->ltd_active = activate; lmv->desc.ld_active_tgt_count += (activate ? 1 : -1); tgt->ltd_exp->exp_obd->obd_inactive = !activate; } /** * Error codes: * * -EINVAL : UUID can't be found in the LMV's target list * -ENOTCONN: The UUID is found, but the target connection is bad (!) * -EBADF : The UUID is found, but the OBD of the wrong type (!) */ static int lmv_set_mdc_active(struct lmv_obd *lmv, const struct obd_uuid *uuid, int activate) { struct lmv_tgt_desc *tgt = NULL; struct obd_device *obd; __u32 i; int rc = 0; ENTRY; CDEBUG(D_INFO, "Searching in lmv %p for uuid %s (activate=%d)\n", lmv, uuid->uuid, activate); spin_lock(&lmv->lmv_lock); for (i = 0; i < lmv->desc.ld_tgt_count; i++) { tgt = lmv->tgts[i]; if (tgt == NULL || tgt->ltd_exp == NULL) continue; CDEBUG(D_INFO, "Target idx %d is %s conn "LPX64"\n", i, tgt->ltd_uuid.uuid, tgt->ltd_exp->exp_handle.h_cookie); if (obd_uuid_equals(uuid, &tgt->ltd_uuid)) break; } if (i == lmv->desc.ld_tgt_count) GOTO(out_lmv_lock, rc = -EINVAL); obd = class_exp2obd(tgt->ltd_exp); if (obd == NULL) GOTO(out_lmv_lock, rc = -ENOTCONN); CDEBUG(D_INFO, "Found OBD %s=%s device %d (%p) type %s at LMV idx %d\n", obd->obd_name, obd->obd_uuid.uuid, obd->obd_minor, obd, obd->obd_type->typ_name, i); LASSERT(strcmp(obd->obd_type->typ_name, LUSTRE_MDC_NAME) == 0); if (tgt->ltd_active == activate) { CDEBUG(D_INFO, "OBD %p already %sactive!\n", obd, activate ? "" : "in"); GOTO(out_lmv_lock, rc); } CDEBUG(D_INFO, "Marking OBD %p %sactive\n", obd, activate ? "" : "in"); lmv_activate_target(lmv, tgt, activate); EXIT; out_lmv_lock: spin_unlock(&lmv->lmv_lock); return rc; } struct obd_uuid *lmv_get_uuid(struct obd_export *exp) { struct lmv_obd *lmv = &exp->exp_obd->u.lmv; struct lmv_tgt_desc *tgt = lmv->tgts[0]; return (tgt == NULL) ? NULL : obd_get_uuid(tgt->ltd_exp); } static int lmv_notify(struct obd_device *obd, struct obd_device *watched, enum obd_notify_event ev, void *data) { struct obd_connect_data *conn_data; struct lmv_obd *lmv = &obd->u.lmv; struct obd_uuid *uuid; int rc = 0; ENTRY; if (strcmp(watched->obd_type->typ_name, LUSTRE_MDC_NAME)) { CERROR("unexpected notification of %s %s!\n", watched->obd_type->typ_name, watched->obd_name); RETURN(-EINVAL); } uuid = &watched->u.cli.cl_target_uuid; if (ev == OBD_NOTIFY_ACTIVE || ev == OBD_NOTIFY_INACTIVE) { /* * Set MDC as active before notifying the observer, so the * observer can use the MDC normally. */ rc = lmv_set_mdc_active(lmv, uuid, ev == OBD_NOTIFY_ACTIVE); if (rc) { CERROR("%sactivation of %s failed: %d\n", ev == OBD_NOTIFY_ACTIVE ? "" : "de", uuid->uuid, rc); RETURN(rc); } } else if (ev == OBD_NOTIFY_OCD) { conn_data = &watched->u.cli.cl_import->imp_connect_data; /* * XXX: Make sure that ocd_connect_flags from all targets are * the same. Otherwise one of MDTs runs wrong version or * something like this. --umka */ obd->obd_self_export->exp_connect_data = *conn_data; } #if 0 else if (ev == OBD_NOTIFY_DISCON) { /* * For disconnect event, flush fld cache for failout MDS case. */ fld_client_flush(&lmv->lmv_fld); } #endif /* * Pass the notification up the chain. */ if (obd->obd_observer) rc = obd_notify(obd->obd_observer, watched, ev, data); RETURN(rc); } /** * This is fake connect function. Its purpose is to initialize lmv and say * caller that everything is okay. Real connection will be performed later. */ static int lmv_connect(const struct lu_env *env, struct obd_export **exp, struct obd_device *obd, struct obd_uuid *cluuid, struct obd_connect_data *data, void *localdata) { struct lmv_obd *lmv = &obd->u.lmv; struct lustre_handle conn = { 0 }; int rc = 0; ENTRY; /* * We don't want to actually do the underlying connections more than * once, so keep track. */ lmv->refcount++; if (lmv->refcount > 1) { *exp = NULL; RETURN(0); } rc = class_connect(&conn, obd, cluuid); if (rc) { CERROR("class_connection() returned %d\n", rc); RETURN(rc); } *exp = class_conn2export(&conn); class_export_get(*exp); lmv->exp = *exp; lmv->connected = 0; lmv->cluuid = *cluuid; if (data) lmv->conn_data = *data; if (lmv->targets_proc_entry == NULL) { lmv->targets_proc_entry = lprocfs_register("target_obds", obd->obd_proc_entry, NULL, NULL); if (IS_ERR(lmv->targets_proc_entry)) { CERROR("%s: cannot register " "/proc/fs/lustre/%s/%s/target_obds\n", obd->obd_name, obd->obd_type->typ_name, obd->obd_name); lmv->targets_proc_entry = NULL; } } /* * All real clients should perform actual connection right away, because * it is possible, that LMV will not have opportunity to connect targets * and MDC stuff will be called directly, for instance while reading * ../mdc/../kbytesfree procfs file, etc. */ if (data != NULL && (data->ocd_connect_flags & OBD_CONNECT_REAL)) rc = lmv_check_connect(obd); if (rc && lmv->targets_proc_entry != NULL) lprocfs_remove(&lmv->targets_proc_entry); RETURN(rc); } static int lmv_init_ea_size(struct obd_export *exp, __u32 easize, __u32 def_easize) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; __u32 i; int rc = 0; int change = 0; ENTRY; if (lmv->max_easize < easize) { lmv->max_easize = easize; change = 1; } if (lmv->max_def_easize < def_easize) { lmv->max_def_easize = def_easize; change = 1; } if (change == 0) RETURN(0); if (lmv->connected == 0) RETURN(0); for (i = 0; i < lmv->desc.ld_tgt_count; i++) { struct lmv_tgt_desc *tgt = lmv->tgts[i]; if (tgt == NULL || tgt->ltd_exp == NULL || !tgt->ltd_active) { CWARN("%s: NULL export for %d\n", obd->obd_name, i); continue; } rc = md_init_ea_size(tgt->ltd_exp, easize, def_easize); if (rc) { CERROR("%s: obd_init_ea_size() failed on MDT target %d:" " rc = %d\n", obd->obd_name, i, rc); break; } } RETURN(rc); } #define MAX_STRING_SIZE 128 int lmv_connect_mdc(struct obd_device *obd, struct lmv_tgt_desc *tgt) { struct lmv_obd *lmv = &obd->u.lmv; struct obd_uuid *cluuid = &lmv->cluuid; struct obd_uuid lmv_mdc_uuid = { "LMV_MDC_UUID" }; struct obd_device *mdc_obd; struct obd_export *mdc_exp; struct lu_fld_target target; int rc; ENTRY; mdc_obd = class_find_client_obd(&tgt->ltd_uuid, LUSTRE_MDC_NAME, &obd->obd_uuid); if (!mdc_obd) { CERROR("target %s not attached\n", tgt->ltd_uuid.uuid); RETURN(-EINVAL); } CDEBUG(D_CONFIG, "connect to %s(%s) - %s, %s FOR %s\n", mdc_obd->obd_name, mdc_obd->obd_uuid.uuid, tgt->ltd_uuid.uuid, obd->obd_uuid.uuid, cluuid->uuid); if (!mdc_obd->obd_set_up) { CERROR("target %s is not set up\n", tgt->ltd_uuid.uuid); RETURN(-EINVAL); } rc = obd_connect(NULL, &mdc_exp, mdc_obd, &lmv_mdc_uuid, &lmv->conn_data, NULL); if (rc) { CERROR("target %s connect error %d\n", tgt->ltd_uuid.uuid, rc); RETURN(rc); } /* * Init fid sequence client for this mdc and add new fld target. */ rc = obd_fid_init(mdc_obd, mdc_exp, LUSTRE_SEQ_METADATA); if (rc) RETURN(rc); target.ft_srv = NULL; target.ft_exp = mdc_exp; target.ft_idx = tgt->ltd_idx; fld_client_add_target(&lmv->lmv_fld, &target); rc = obd_register_observer(mdc_obd, obd); if (rc) { obd_disconnect(mdc_exp); CERROR("target %s register_observer error %d\n", tgt->ltd_uuid.uuid, rc); RETURN(rc); } if (obd->obd_observer) { /* * Tell the observer about the new target. */ rc = obd_notify(obd->obd_observer, mdc_exp->exp_obd, OBD_NOTIFY_ACTIVE, (void *)(tgt - lmv->tgts[0])); if (rc) { obd_disconnect(mdc_exp); RETURN(rc); } } tgt->ltd_active = 1; tgt->ltd_exp = mdc_exp; lmv->desc.ld_active_tgt_count++; md_init_ea_size(tgt->ltd_exp, lmv->max_easize, lmv->max_def_easize); CDEBUG(D_CONFIG, "Connected to %s(%s) successfully (%d)\n", mdc_obd->obd_name, mdc_obd->obd_uuid.uuid, atomic_read(&obd->obd_refcount)); if (lmv->targets_proc_entry != NULL) { struct proc_dir_entry *mdc_symlink; LASSERT(mdc_obd->obd_type != NULL); LASSERT(mdc_obd->obd_type->typ_name != NULL); mdc_symlink = lprocfs_add_symlink(mdc_obd->obd_name, lmv->targets_proc_entry, "../../../%s/%s", mdc_obd->obd_type->typ_name, mdc_obd->obd_name); if (mdc_symlink == NULL) { CERROR("cannot register LMV target " "/proc/fs/lustre/%s/%s/target_obds/%s\n", obd->obd_type->typ_name, obd->obd_name, mdc_obd->obd_name); } } RETURN(0); } static void lmv_del_target(struct lmv_obd *lmv, int index) { if (lmv->tgts[index] == NULL) return; OBD_FREE_PTR(lmv->tgts[index]); lmv->tgts[index] = NULL; return; } static int lmv_add_target(struct obd_device *obd, struct obd_uuid *uuidp, __u32 index, int gen) { struct obd_device *mdc_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int orig_tgt_count = 0; int rc = 0; ENTRY; CDEBUG(D_CONFIG, "Target uuid: %s. index %d\n", uuidp->uuid, index); mdc_obd = class_find_client_obd(uuidp, LUSTRE_MDC_NAME, &obd->obd_uuid); if (!mdc_obd) { CERROR("%s: Target %s not attached: rc = %d\n", obd->obd_name, uuidp->uuid, -EINVAL); RETURN(-EINVAL); } mutex_lock(&lmv->lmv_init_mutex); if ((index < lmv->tgts_size) && (lmv->tgts[index] != NULL)) { tgt = lmv->tgts[index]; CERROR("%s: UUID %s already assigned at LOV target index %d:" " rc = %d\n", obd->obd_name, obd_uuid2str(&tgt->ltd_uuid), index, -EEXIST); mutex_unlock(&lmv->lmv_init_mutex); RETURN(-EEXIST); } if (index >= lmv->tgts_size) { /* We need to reallocate the lmv target array. */ struct lmv_tgt_desc **newtgts, **old = NULL; __u32 newsize = 1; __u32 oldsize = 0; while (newsize < index + 1) newsize = newsize << 1; OBD_ALLOC(newtgts, sizeof(*newtgts) * newsize); if (newtgts == NULL) { mutex_unlock(&lmv->lmv_init_mutex); RETURN(-ENOMEM); } if (lmv->tgts_size) { memcpy(newtgts, lmv->tgts, sizeof(*newtgts) * lmv->tgts_size); old = lmv->tgts; oldsize = lmv->tgts_size; } lmv->tgts = newtgts; lmv->tgts_size = newsize; smp_rmb(); if (old) OBD_FREE(old, sizeof(*old) * oldsize); CDEBUG(D_CONFIG, "tgts: %p size: %d\n", lmv->tgts, lmv->tgts_size); } OBD_ALLOC_PTR(tgt); if (!tgt) { mutex_unlock(&lmv->lmv_init_mutex); RETURN(-ENOMEM); } mutex_init(&tgt->ltd_fid_mutex); tgt->ltd_idx = index; tgt->ltd_uuid = *uuidp; tgt->ltd_active = 0; lmv->tgts[index] = tgt; if (index >= lmv->desc.ld_tgt_count) { orig_tgt_count = lmv->desc.ld_tgt_count; lmv->desc.ld_tgt_count = index + 1; } if (lmv->connected == 0) { /* lmv_check_connect() will connect this target. */ mutex_unlock(&lmv->lmv_init_mutex); RETURN(0); } /* Otherwise let's connect it ourselves */ mutex_unlock(&lmv->lmv_init_mutex); rc = lmv_connect_mdc(obd, tgt); if (rc != 0) { spin_lock(&lmv->lmv_lock); if (lmv->desc.ld_tgt_count == index + 1) lmv->desc.ld_tgt_count = orig_tgt_count; memset(tgt, 0, sizeof(*tgt)); spin_unlock(&lmv->lmv_lock); } else { int easize = sizeof(struct lmv_stripe_md) + lmv->desc.ld_tgt_count * sizeof(struct lu_fid); lmv_init_ea_size(obd->obd_self_export, easize, 0); } RETURN(rc); } int lmv_check_connect(struct obd_device *obd) { struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; __u32 i; int rc; int easize; ENTRY; if (lmv->connected) RETURN(0); mutex_lock(&lmv->lmv_init_mutex); if (lmv->connected) { mutex_unlock(&lmv->lmv_init_mutex); RETURN(0); } if (lmv->desc.ld_tgt_count == 0) { mutex_unlock(&lmv->lmv_init_mutex); CERROR("%s: no targets configured.\n", obd->obd_name); RETURN(-EINVAL); } LASSERT(lmv->tgts != NULL); if (lmv->tgts[0] == NULL) { mutex_unlock(&lmv->lmv_init_mutex); CERROR("%s: no target configured for index 0.\n", obd->obd_name); RETURN(-EINVAL); } CDEBUG(D_CONFIG, "Time to connect %s to %s\n", lmv->cluuid.uuid, obd->obd_name); for (i = 0; i < lmv->desc.ld_tgt_count; i++) { tgt = lmv->tgts[i]; if (tgt == NULL) continue; rc = lmv_connect_mdc(obd, tgt); if (rc) GOTO(out_disc, rc); } class_export_put(lmv->exp); lmv->connected = 1; easize = lmv_mds_md_size(lmv->desc.ld_tgt_count, LMV_MAGIC); lmv_init_ea_size(obd->obd_self_export, easize, 0); mutex_unlock(&lmv->lmv_init_mutex); RETURN(0); out_disc: while (i-- > 0) { int rc2; tgt = lmv->tgts[i]; if (tgt == NULL) continue; tgt->ltd_active = 0; if (tgt->ltd_exp) { --lmv->desc.ld_active_tgt_count; rc2 = obd_disconnect(tgt->ltd_exp); if (rc2) { CERROR("LMV target %s disconnect on " "MDC idx %d: error %d\n", tgt->ltd_uuid.uuid, i, rc2); } } } class_disconnect(lmv->exp); mutex_unlock(&lmv->lmv_init_mutex); RETURN(rc); } static int lmv_disconnect_mdc(struct obd_device *obd, struct lmv_tgt_desc *tgt) { struct lmv_obd *lmv = &obd->u.lmv; struct obd_device *mdc_obd; int rc; ENTRY; LASSERT(tgt != NULL); LASSERT(obd != NULL); mdc_obd = class_exp2obd(tgt->ltd_exp); if (mdc_obd) { mdc_obd->obd_force = obd->obd_force; mdc_obd->obd_fail = obd->obd_fail; mdc_obd->obd_no_recov = obd->obd_no_recov; if (lmv->targets_proc_entry != NULL) lprocfs_remove_proc_entry(mdc_obd->obd_name, lmv->targets_proc_entry); } rc = obd_fid_fini(tgt->ltd_exp->exp_obd); if (rc) CERROR("Can't finanize fids factory\n"); CDEBUG(D_INFO, "Disconnected from %s(%s) successfully\n", tgt->ltd_exp->exp_obd->obd_name, tgt->ltd_exp->exp_obd->obd_uuid.uuid); obd_register_observer(tgt->ltd_exp->exp_obd, NULL); rc = obd_disconnect(tgt->ltd_exp); if (rc) { if (tgt->ltd_active) { CERROR("Target %s disconnect error %d\n", tgt->ltd_uuid.uuid, rc); } } lmv_activate_target(lmv, tgt, 0); tgt->ltd_exp = NULL; RETURN(0); } static int lmv_disconnect(struct obd_export *exp) { struct obd_device *obd = class_exp2obd(exp); struct lmv_obd *lmv = &obd->u.lmv; int rc; __u32 i; ENTRY; if (!lmv->tgts) goto out_local; /* * Only disconnect the underlying layers on the final disconnect. */ lmv->refcount--; if (lmv->refcount != 0) goto out_local; for (i = 0; i < lmv->desc.ld_tgt_count; i++) { if (lmv->tgts[i] == NULL || lmv->tgts[i]->ltd_exp == NULL) continue; lmv_disconnect_mdc(obd, lmv->tgts[i]); } if (lmv->targets_proc_entry != NULL) lprocfs_remove(&lmv->targets_proc_entry); else CERROR("/proc/fs/lustre/%s/%s/target_obds missing\n", obd->obd_type->typ_name, obd->obd_name); out_local: /* * This is the case when no real connection is established by * lmv_check_connect(). */ if (!lmv->connected) class_export_put(exp); rc = class_disconnect(exp); if (lmv->refcount == 0) lmv->connected = 0; RETURN(rc); } static int lmv_fid2path(struct obd_export *exp, int len, void *karg, void __user *uarg) { struct obd_device *obddev = class_exp2obd(exp); struct lmv_obd *lmv = &obddev->u.lmv; struct getinfo_fid2path *gf; struct lmv_tgt_desc *tgt; struct getinfo_fid2path *remote_gf = NULL; int remote_gf_size = 0; int rc; gf = (struct getinfo_fid2path *)karg; tgt = lmv_find_target(lmv, &gf->gf_fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); repeat_fid2path: rc = obd_iocontrol(OBD_IOC_FID2PATH, tgt->ltd_exp, len, gf, uarg); if (rc != 0 && rc != -EREMOTE) GOTO(out_fid2path, rc); /* If remote_gf != NULL, it means just building the * path on the remote MDT, copy this path segement to gf */ if (remote_gf != NULL) { struct getinfo_fid2path *ori_gf; char *ptr; ori_gf = (struct getinfo_fid2path *)karg; if (strlen(ori_gf->gf_path) + strlen(gf->gf_path) > ori_gf->gf_pathlen) GOTO(out_fid2path, rc = -EOVERFLOW); ptr = ori_gf->gf_path; memmove(ptr + strlen(gf->gf_path) + 1, ptr, strlen(ori_gf->gf_path)); strncpy(ptr, gf->gf_path, strlen(gf->gf_path)); ptr += strlen(gf->gf_path); *ptr = '/'; } CDEBUG(D_INFO, "%s: get path %s "DFID" rec: "LPU64" ln: %u\n", tgt->ltd_exp->exp_obd->obd_name, gf->gf_path, PFID(&gf->gf_fid), gf->gf_recno, gf->gf_linkno); if (rc == 0) GOTO(out_fid2path, rc); /* sigh, has to go to another MDT to do path building further */ if (remote_gf == NULL) { remote_gf_size = sizeof(*remote_gf) + PATH_MAX; OBD_ALLOC(remote_gf, remote_gf_size); if (remote_gf == NULL) GOTO(out_fid2path, rc = -ENOMEM); remote_gf->gf_pathlen = PATH_MAX; } if (!fid_is_sane(&gf->gf_fid)) { CERROR("%s: invalid FID "DFID": rc = %d\n", tgt->ltd_exp->exp_obd->obd_name, PFID(&gf->gf_fid), -EINVAL); GOTO(out_fid2path, rc = -EINVAL); } tgt = lmv_find_target(lmv, &gf->gf_fid); if (IS_ERR(tgt)) GOTO(out_fid2path, rc = -EINVAL); remote_gf->gf_fid = gf->gf_fid; remote_gf->gf_recno = -1; remote_gf->gf_linkno = -1; memset(remote_gf->gf_path, 0, remote_gf->gf_pathlen); gf = remote_gf; goto repeat_fid2path; out_fid2path: if (remote_gf != NULL) OBD_FREE(remote_gf, remote_gf_size); RETURN(rc); } static int lmv_hsm_req_count(struct lmv_obd *lmv, const struct hsm_user_request *hur, const struct lmv_tgt_desc *tgt_mds) { __u32 i; int nr = 0; struct lmv_tgt_desc *curr_tgt; /* count how many requests must be sent to the given target */ for (i = 0; i < hur->hur_request.hr_itemcount; i++) { curr_tgt = lmv_find_target(lmv, &hur->hur_user_item[i].hui_fid); if (obd_uuid_equals(&curr_tgt->ltd_uuid, &tgt_mds->ltd_uuid)) nr++; } return nr; } static void lmv_hsm_req_build(struct lmv_obd *lmv, struct hsm_user_request *hur_in, const struct lmv_tgt_desc *tgt_mds, struct hsm_user_request *hur_out) { __u32 i, nr_out; struct lmv_tgt_desc *curr_tgt; /* build the hsm_user_request for the given target */ hur_out->hur_request = hur_in->hur_request; nr_out = 0; for (i = 0; i < hur_in->hur_request.hr_itemcount; i++) { curr_tgt = lmv_find_target(lmv, &hur_in->hur_user_item[i].hui_fid); if (obd_uuid_equals(&curr_tgt->ltd_uuid, &tgt_mds->ltd_uuid)) { hur_out->hur_user_item[nr_out] = hur_in->hur_user_item[i]; nr_out++; } } hur_out->hur_request.hr_itemcount = nr_out; memcpy(hur_data(hur_out), hur_data(hur_in), hur_in->hur_request.hr_data_len); } static int lmv_hsm_ct_unregister(struct lmv_obd *lmv, unsigned int cmd, int len, struct lustre_kernelcomm *lk, void *uarg) { __u32 i; int rc; ENTRY; /* unregister request (call from llapi_hsm_copytool_fini) */ for (i = 0; i < lmv->desc.ld_tgt_count; i++) { struct lmv_tgt_desc *tgt = lmv->tgts[i]; if (tgt == NULL || tgt->ltd_exp == NULL) continue; /* best effort: try to clean as much as possible * (continue on error) */ obd_iocontrol(cmd, tgt->ltd_exp, len, lk, uarg); } /* Whatever the result, remove copytool from kuc groups. * Unreached coordinators will get EPIPE on next requests * and will unregister automatically. */ rc = libcfs_kkuc_group_rem(lk->lk_uid, lk->lk_group); RETURN(rc); } static int lmv_hsm_ct_register(struct lmv_obd *lmv, unsigned int cmd, int len, struct lustre_kernelcomm *lk, void *uarg) { struct file *filp; __u32 i, j; int err, rc; bool any_set = false; struct kkuc_ct_data kcd = { 0 }; ENTRY; /* All or nothing: try to register to all MDS. * In case of failure, unregister from previous MDS, * except if it because of inactive target. */ for (i = 0; i < lmv->desc.ld_tgt_count; i++) { struct lmv_tgt_desc *tgt = lmv->tgts[i]; if (tgt == NULL || tgt->ltd_exp == NULL) continue; err = obd_iocontrol(cmd, tgt->ltd_exp, len, lk, uarg); if (err) { if (tgt->ltd_active) { /* permanent error */ CERROR("%s: iocontrol MDC %s on MDT" " idx %d cmd %x: err = %d\n", class_exp2obd(lmv->exp)->obd_name, tgt->ltd_uuid.uuid, i, cmd, err); rc = err; lk->lk_flags |= LK_FLG_STOP; /* unregister from previous MDS */ for (j = 0; j < i; j++) { tgt = lmv->tgts[j]; if (tgt == NULL || tgt->ltd_exp == NULL) continue; obd_iocontrol(cmd, tgt->ltd_exp, len, lk, uarg); } RETURN(rc); } /* else: transient error. * kuc will register to the missing MDT * when it is back */ } else { any_set = true; } } if (!any_set) /* no registration done: return error */ RETURN(-ENOTCONN); /* at least one registration done, with no failure */ filp = fget(lk->lk_wfd); if (filp == NULL) RETURN(-EBADF); kcd.kcd_magic = KKUC_CT_DATA_MAGIC; kcd.kcd_uuid = lmv->cluuid; kcd.kcd_archive = lk->lk_data; rc = libcfs_kkuc_group_add(filp, lk->lk_uid, lk->lk_group, &kcd, sizeof(kcd)); if (rc != 0) fput(filp); RETURN(rc); } static int lmv_iocontrol(unsigned int cmd, struct obd_export *exp, int len, void *karg, void __user *uarg) { struct obd_device *obddev = class_exp2obd(exp); struct lmv_obd *lmv = &obddev->u.lmv; struct lmv_tgt_desc *tgt = NULL; __u32 i = 0; int rc = 0; int set = 0; __u32 count = lmv->desc.ld_tgt_count; ENTRY; if (count == 0) RETURN(-ENOTTY); switch (cmd) { case IOC_OBD_STATFS: { struct obd_ioctl_data *data = karg; struct obd_device *mdc_obd; struct obd_statfs stat_buf = {0}; __u32 index; memcpy(&index, data->ioc_inlbuf2, sizeof(__u32)); if ((index >= count)) RETURN(-ENODEV); tgt = lmv->tgts[index]; if (tgt == NULL || !tgt->ltd_active) RETURN(-ENODATA); mdc_obd = class_exp2obd(tgt->ltd_exp); if (!mdc_obd) RETURN(-EINVAL); /* copy UUID */ if (copy_to_user(data->ioc_pbuf2, obd2cli_tgt(mdc_obd), min((int) data->ioc_plen2, (int) sizeof(struct obd_uuid)))) RETURN(-EFAULT); rc = obd_statfs(NULL, tgt->ltd_exp, &stat_buf, cfs_time_shift_64(-OBD_STATFS_CACHE_SECONDS), 0); if (rc) RETURN(rc); if (copy_to_user(data->ioc_pbuf1, &stat_buf, min((int) data->ioc_plen1, (int) sizeof(stat_buf)))) RETURN(-EFAULT); break; } case OBD_IOC_QUOTACTL: { struct if_quotactl *qctl = karg; struct obd_quotactl *oqctl; if (qctl->qc_valid == QC_MDTIDX) { if (count <= qctl->qc_idx) RETURN(-EINVAL); tgt = lmv->tgts[qctl->qc_idx]; if (tgt == NULL || tgt->ltd_exp == NULL) RETURN(-EINVAL); } else if (qctl->qc_valid == QC_UUID) { for (i = 0; i < count; i++) { tgt = lmv->tgts[i]; if (tgt == NULL) continue; if (!obd_uuid_equals(&tgt->ltd_uuid, &qctl->obd_uuid)) continue; if (tgt->ltd_exp == NULL) RETURN(-EINVAL); break; } } else { RETURN(-EINVAL); } if (i >= count) RETURN(-EAGAIN); LASSERT(tgt != NULL && tgt->ltd_exp != NULL); OBD_ALLOC_PTR(oqctl); if (!oqctl) RETURN(-ENOMEM); QCTL_COPY(oqctl, qctl); rc = obd_quotactl(tgt->ltd_exp, oqctl); if (rc == 0) { QCTL_COPY(qctl, oqctl); qctl->qc_valid = QC_MDTIDX; qctl->obd_uuid = tgt->ltd_uuid; } OBD_FREE_PTR(oqctl); break; } case OBD_IOC_CHANGELOG_SEND: case OBD_IOC_CHANGELOG_CLEAR: { struct ioc_changelog *icc = karg; if (icc->icc_mdtindex >= count) RETURN(-ENODEV); tgt = lmv->tgts[icc->icc_mdtindex]; if (tgt == NULL || tgt->ltd_exp == NULL || !tgt->ltd_active) RETURN(-ENODEV); rc = obd_iocontrol(cmd, tgt->ltd_exp, sizeof(*icc), icc, NULL); break; } case LL_IOC_GET_CONNECT_FLAGS: { tgt = lmv->tgts[0]; if (tgt == NULL || tgt->ltd_exp == NULL) RETURN(-ENODATA); rc = obd_iocontrol(cmd, tgt->ltd_exp, len, karg, uarg); break; } case LL_IOC_FID2MDTIDX: { struct lu_fid *fid = karg; int mdt_index; rc = lmv_fld_lookup(lmv, fid, &mdt_index); if (rc != 0) RETURN(rc); /* Note: this is from llite(see ll_dir_ioctl()), @uarg does not * point to user space memory for FID2MDTIDX. */ *(__u32 *)uarg = mdt_index; break; } case OBD_IOC_FID2PATH: { rc = lmv_fid2path(exp, len, karg, uarg); break; } case LL_IOC_HSM_STATE_GET: case LL_IOC_HSM_STATE_SET: case LL_IOC_HSM_ACTION: { struct md_op_data *op_data = karg; tgt = lmv_find_target(lmv, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); if (tgt->ltd_exp == NULL) RETURN(-EINVAL); rc = obd_iocontrol(cmd, tgt->ltd_exp, len, karg, uarg); break; } case LL_IOC_HSM_PROGRESS: { const struct hsm_progress_kernel *hpk = karg; tgt = lmv_find_target(lmv, &hpk->hpk_fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); rc = obd_iocontrol(cmd, tgt->ltd_exp, len, karg, uarg); break; } case LL_IOC_HSM_REQUEST: { struct hsm_user_request *hur = karg; unsigned int reqcount = hur->hur_request.hr_itemcount; if (reqcount == 0) RETURN(0); /* if the request is about a single fid * or if there is a single MDS, no need to split * the request. */ if (reqcount == 1 || count == 1) { tgt = lmv_find_target(lmv, &hur->hur_user_item[0].hui_fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); rc = obd_iocontrol(cmd, tgt->ltd_exp, len, karg, uarg); } else { /* split fid list to their respective MDS */ for (i = 0; i < count; i++) { unsigned int nr, reqlen; int rc1; struct hsm_user_request *req; tgt = lmv->tgts[i]; if (tgt == NULL || tgt->ltd_exp == NULL) continue; nr = lmv_hsm_req_count(lmv, hur, tgt); if (nr == 0) /* nothing for this MDS */ continue; /* build a request with fids for this MDS */ reqlen = offsetof(typeof(*hur), hur_user_item[nr]) + hur->hur_request.hr_data_len; OBD_ALLOC_LARGE(req, reqlen); if (req == NULL) RETURN(-ENOMEM); lmv_hsm_req_build(lmv, hur, tgt, req); rc1 = obd_iocontrol(cmd, tgt->ltd_exp, reqlen, req, uarg); if (rc1 != 0 && rc == 0) rc = rc1; OBD_FREE_LARGE(req, reqlen); } } break; } case LL_IOC_LOV_SWAP_LAYOUTS: { struct md_op_data *op_data = karg; struct lmv_tgt_desc *tgt1, *tgt2; tgt1 = lmv_find_target(lmv, &op_data->op_fid1); if (IS_ERR(tgt1)) RETURN(PTR_ERR(tgt1)); tgt2 = lmv_find_target(lmv, &op_data->op_fid2); if (IS_ERR(tgt2)) RETURN(PTR_ERR(tgt2)); if ((tgt1->ltd_exp == NULL) || (tgt2->ltd_exp == NULL)) RETURN(-EINVAL); /* only files on same MDT can have their layouts swapped */ if (tgt1->ltd_idx != tgt2->ltd_idx) RETURN(-EPERM); rc = obd_iocontrol(cmd, tgt1->ltd_exp, len, karg, uarg); break; } case LL_IOC_HSM_CT_START: { struct lustre_kernelcomm *lk = karg; if (lk->lk_flags & LK_FLG_STOP) rc = lmv_hsm_ct_unregister(lmv, cmd, len, lk, uarg); else rc = lmv_hsm_ct_register(lmv, cmd, len, lk, uarg); break; } default: for (i = 0; i < count; i++) { struct obd_device *mdc_obd; int err; tgt = lmv->tgts[i]; if (tgt == NULL || tgt->ltd_exp == NULL) continue; /* ll_umount_begin() sets force flag but for lmv, not * mdc. Let's pass it through */ mdc_obd = class_exp2obd(tgt->ltd_exp); mdc_obd->obd_force = obddev->obd_force; err = obd_iocontrol(cmd, tgt->ltd_exp, len, karg, uarg); if (err) { if (tgt->ltd_active) { CERROR("error: iocontrol MDC %s on MDT" " idx %d cmd %x: err = %d\n", tgt->ltd_uuid.uuid, i, cmd, err); if (!rc) rc = err; } } else set = 1; } if (!set && !rc) rc = -EIO; } RETURN(rc); } #if 0 static int lmv_all_chars_policy(int count, const char *name, int len) { unsigned int c = 0; while (len > 0) c += name[--len]; c = c % count; return c; } static int lmv_nid_policy(struct lmv_obd *lmv) { struct obd_import *imp; __u32 id; /* * XXX: To get nid we assume that underlying obd device is mdc. */ imp = class_exp2cliimp(lmv->tgts[0].ltd_exp); id = imp->imp_connection->c_self ^ (imp->imp_connection->c_self >> 32); return id % lmv->desc.ld_tgt_count; } static int lmv_choose_mds(struct lmv_obd *lmv, struct md_op_data *op_data, placement_policy_t placement) { switch (placement) { case PLACEMENT_CHAR_POLICY: return lmv_all_chars_policy(lmv->desc.ld_tgt_count, op_data->op_name, op_data->op_namelen); case PLACEMENT_NID_POLICY: return lmv_nid_policy(lmv); default: break; } CERROR("Unsupported placement policy %x\n", placement); return -EINVAL; } #endif /** * This is _inode_ placement policy function (not name). */ static int lmv_placement_policy(struct obd_device *obd, struct md_op_data *op_data, u32 *mds) { struct lmv_obd *lmv = &obd->u.lmv; ENTRY; LASSERT(mds != NULL); if (lmv->desc.ld_tgt_count == 1) { *mds = 0; RETURN(0); } if (op_data->op_default_stripe_offset != -1) { *mds = op_data->op_default_stripe_offset; RETURN(0); } /** * If stripe_offset is provided during setdirstripe * (setdirstripe -i xx), xx MDS will be choosen. */ if (op_data->op_cli_flags & CLI_SET_MEA && op_data->op_data != NULL) { struct lmv_user_md *lum; lum = op_data->op_data; if (le32_to_cpu(lum->lum_stripe_offset) != (__u32)-1) { *mds = le32_to_cpu(lum->lum_stripe_offset); } else { /* -1 means default, which will be in the same MDT with * the stripe */ *mds = op_data->op_mds; lum->lum_stripe_offset = cpu_to_le32(op_data->op_mds); } } else { /* Allocate new fid on target according to operation type and * parent home mds. */ *mds = op_data->op_mds; } RETURN(0); } int __lmv_fid_alloc(struct lmv_obd *lmv, struct lu_fid *fid, u32 mds) { struct lmv_tgt_desc *tgt; int rc; ENTRY; tgt = lmv_get_target(lmv, mds, NULL); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); /* * New seq alloc and FLD setup should be atomic. Otherwise we may find * on server that seq in new allocated fid is not yet known. */ mutex_lock(&tgt->ltd_fid_mutex); if (tgt->ltd_active == 0 || tgt->ltd_exp == NULL) GOTO(out, rc = -ENODEV); /* * Asking underlying tgt layer to allocate new fid. */ rc = obd_fid_alloc(NULL, tgt->ltd_exp, fid, NULL); if (rc > 0) { LASSERT(fid_is_sane(fid)); rc = 0; } EXIT; out: mutex_unlock(&tgt->ltd_fid_mutex); return rc; } int lmv_fid_alloc(const struct lu_env *env, struct obd_export *exp, struct lu_fid *fid, struct md_op_data *op_data) { struct obd_device *obd = class_exp2obd(exp); struct lmv_obd *lmv = &obd->u.lmv; u32 mds = 0; int rc; ENTRY; LASSERT(op_data != NULL); LASSERT(fid != NULL); rc = lmv_placement_policy(obd, op_data, &mds); if (rc) { CERROR("Can't get target for allocating fid, " "rc %d\n", rc); RETURN(rc); } rc = __lmv_fid_alloc(lmv, fid, mds); if (rc) { CERROR("Can't alloc new fid, rc %d\n", rc); RETURN(rc); } RETURN(rc); } static int lmv_setup(struct obd_device *obd, struct lustre_cfg *lcfg) { struct lmv_obd *lmv = &obd->u.lmv; struct lmv_desc *desc; int rc; ENTRY; if (LUSTRE_CFG_BUFLEN(lcfg, 1) < 1) { CERROR("LMV setup requires a descriptor\n"); RETURN(-EINVAL); } desc = (struct lmv_desc *)lustre_cfg_buf(lcfg, 1); if (sizeof(*desc) > LUSTRE_CFG_BUFLEN(lcfg, 1)) { CERROR("Lmv descriptor size wrong: %d > %d\n", (int)sizeof(*desc), LUSTRE_CFG_BUFLEN(lcfg, 1)); RETURN(-EINVAL); } lmv->tgts_size = 32U; OBD_ALLOC(lmv->tgts, sizeof(*lmv->tgts) * lmv->tgts_size); if (lmv->tgts == NULL) RETURN(-ENOMEM); obd_str2uuid(&lmv->desc.ld_uuid, desc->ld_uuid.uuid); lmv->desc.ld_tgt_count = 0; lmv->desc.ld_active_tgt_count = 0; lmv->max_def_easize = 0; lmv->max_easize = 0; lmv->lmv_placement = PLACEMENT_CHAR_POLICY; spin_lock_init(&lmv->lmv_lock); mutex_init(&lmv->lmv_init_mutex); #ifdef CONFIG_PROC_FS obd->obd_vars = lprocfs_lmv_obd_vars; lprocfs_obd_setup(obd); lprocfs_alloc_md_stats(obd, 0); rc = lprocfs_seq_create(obd->obd_proc_entry, "target_obd", 0444, &lmv_proc_target_fops, obd); if (rc) CWARN("%s: error adding LMV target_obd file: rc = %d\n", obd->obd_name, rc); #endif rc = fld_client_init(&lmv->lmv_fld, obd->obd_name, LUSTRE_CLI_FLD_HASH_DHT); if (rc) { CERROR("Can't init FLD, err %d\n", rc); GOTO(out, rc); } RETURN(0); out: return rc; } static int lmv_cleanup(struct obd_device *obd) { struct lmv_obd *lmv = &obd->u.lmv; ENTRY; fld_client_fini(&lmv->lmv_fld); if (lmv->tgts != NULL) { int i; for (i = 0; i < lmv->desc.ld_tgt_count; i++) { if (lmv->tgts[i] == NULL) continue; lmv_del_target(lmv, i); } OBD_FREE(lmv->tgts, sizeof(*lmv->tgts) * lmv->tgts_size); lmv->tgts_size = 0; } RETURN(0); } static int lmv_process_config(struct obd_device *obd, size_t len, void *buf) { struct lustre_cfg *lcfg = buf; struct obd_uuid obd_uuid; int gen; __u32 index; int rc; ENTRY; switch (lcfg->lcfg_command) { case LCFG_ADD_MDC: /* modify_mdc_tgts add 0:lustre-clilmv 1:lustre-MDT0000_UUID * 2:0 3:1 4:lustre-MDT0000-mdc_UUID */ if (LUSTRE_CFG_BUFLEN(lcfg, 1) > sizeof(obd_uuid.uuid)) GOTO(out, rc = -EINVAL); obd_str2uuid(&obd_uuid, lustre_cfg_buf(lcfg, 1)); if (sscanf(lustre_cfg_buf(lcfg, 2), "%u", &index) != 1) GOTO(out, rc = -EINVAL); if (sscanf(lustre_cfg_buf(lcfg, 3), "%d", &gen) != 1) GOTO(out, rc = -EINVAL); rc = lmv_add_target(obd, &obd_uuid, index, gen); GOTO(out, rc); default: CERROR("Unknown command: %d\n", lcfg->lcfg_command); GOTO(out, rc = -EINVAL); } out: RETURN(rc); } static int lmv_statfs(const struct lu_env *env, struct obd_export *exp, struct obd_statfs *osfs, __u64 max_age, __u32 flags) { struct obd_device *obd = class_exp2obd(exp); struct lmv_obd *lmv = &obd->u.lmv; struct obd_statfs *temp; int rc = 0; __u32 i; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); OBD_ALLOC(temp, sizeof(*temp)); if (temp == NULL) RETURN(-ENOMEM); for (i = 0; i < lmv->desc.ld_tgt_count; i++) { if (lmv->tgts[i] == NULL || lmv->tgts[i]->ltd_exp == NULL) continue; rc = obd_statfs(env, lmv->tgts[i]->ltd_exp, temp, max_age, flags); if (rc) { CERROR("can't stat MDS #%d (%s), error %d\n", i, lmv->tgts[i]->ltd_exp->exp_obd->obd_name, rc); GOTO(out_free_temp, rc); } if (i == 0) { *osfs = *temp; /* If the statfs is from mount, it will needs * retrieve necessary information from MDT0. * i.e. mount does not need the merged osfs * from all of MDT. * And also clients can be mounted as long as * MDT0 is in service*/ if (flags & OBD_STATFS_FOR_MDT0) GOTO(out_free_temp, rc); } else { osfs->os_bavail += temp->os_bavail; osfs->os_blocks += temp->os_blocks; osfs->os_ffree += temp->os_ffree; osfs->os_files += temp->os_files; } } EXIT; out_free_temp: OBD_FREE(temp, sizeof(*temp)); return rc; } static int lmv_getstatus(struct obd_export *exp, struct lu_fid *fid) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); rc = md_getstatus(lmv->tgts[0]->ltd_exp, fid); RETURN(rc); } static int lmv_getxattr(struct obd_export *exp, const struct lu_fid *fid, u64 valid, const char *name, const char *input, int input_size, int output_size, int flags, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); tgt = lmv_find_target(lmv, fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); rc = md_getxattr(tgt->ltd_exp, fid, valid, name, input, input_size, output_size, flags, request); RETURN(rc); } static int lmv_setxattr(struct obd_export *exp, const struct lu_fid *fid, u64 valid, const char *name, const char *input, int input_size, int output_size, int flags, __u32 suppgid, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); tgt = lmv_find_target(lmv, fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); rc = md_setxattr(tgt->ltd_exp, fid, valid, name, input, input_size, output_size, flags, suppgid, request); RETURN(rc); } static int lmv_getattr(struct obd_export *exp, struct md_op_data *op_data, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); tgt = lmv_find_target(lmv, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); if (op_data->op_flags & MF_GET_MDT_IDX) { op_data->op_mds = tgt->ltd_idx; RETURN(0); } rc = md_getattr(tgt->ltd_exp, op_data, request); RETURN(rc); } static int lmv_null_inode(struct obd_export *exp, const struct lu_fid *fid) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; __u32 i; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); CDEBUG(D_INODE, "CBDATA for "DFID"\n", PFID(fid)); /* * With DNE every object can have two locks in different namespaces: * lookup lock in space of MDT storing direntry and update/open lock in * space of MDT storing inode. */ for (i = 0; i < lmv->desc.ld_tgt_count; i++) { if (lmv->tgts[i] == NULL || lmv->tgts[i]->ltd_exp == NULL) continue; md_null_inode(lmv->tgts[i]->ltd_exp, fid); } RETURN(0); } static int lmv_find_cbdata(struct obd_export *exp, const struct lu_fid *fid, ldlm_iterator_t it, void *data) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; int i; int tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); CDEBUG(D_INODE, "CBDATA for "DFID"\n", PFID(fid)); /* * With DNE every object can have two locks in different namespaces: * lookup lock in space of MDT storing direntry and update/open lock in * space of MDT storing inode. Try the MDT that the FID maps to first, * since this can be easily found, and only try others if that fails. */ for (i = 0, tgt = lmv_find_target_index(lmv, fid); i < lmv->desc.ld_tgt_count; i++, tgt = (tgt + 1) % lmv->desc.ld_tgt_count) { if (tgt < 0) { CDEBUG(D_HA, "%s: "DFID" is inaccessible: rc = %d\n", obd->obd_name, PFID(fid), tgt); tgt = 0; } if (lmv->tgts[tgt] == NULL || lmv->tgts[tgt]->ltd_exp == NULL) continue; rc = md_find_cbdata(lmv->tgts[tgt]->ltd_exp, fid, it, data); if (rc) RETURN(rc); } RETURN(rc); } static int lmv_close(struct obd_export *exp, struct md_op_data *op_data, struct md_open_data *mod, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); tgt = lmv_find_target(lmv, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); CDEBUG(D_INODE, "CLOSE "DFID"\n", PFID(&op_data->op_fid1)); rc = md_close(tgt->ltd_exp, op_data, mod, request); RETURN(rc); } /** * Choosing the MDT by name or FID in @op_data. * For non-striped directory, it will locate MDT by fid. * For striped-directory, it will locate MDT by name. And also * it will reset op_fid1 with the FID of the choosen stripe. **/ struct lmv_tgt_desc * lmv_locate_target_for_name(struct lmv_obd *lmv, struct lmv_stripe_md *lsm, const char *name, int namelen, struct lu_fid *fid, u32 *mds) { struct lmv_tgt_desc *tgt; const struct lmv_oinfo *oinfo; if (OBD_FAIL_CHECK(OBD_FAIL_LFSCK_BAD_NAME_HASH)) { if (cfs_fail_val >= lsm->lsm_md_stripe_count) RETURN(ERR_PTR(-EBADF)); oinfo = &lsm->lsm_md_oinfo[cfs_fail_val]; } else { oinfo = lsm_name_to_stripe_info(lsm, name, namelen); if (IS_ERR(oinfo)) RETURN(ERR_CAST(oinfo)); } if (fid != NULL) *fid = oinfo->lmo_fid; if (mds != NULL) *mds = oinfo->lmo_mds; tgt = lmv_get_target(lmv, oinfo->lmo_mds, NULL); CDEBUG(D_INFO, "locate on mds %u "DFID"\n", oinfo->lmo_mds, PFID(&oinfo->lmo_fid)); return tgt; } /** * Locate mds by fid or name * * For striped directory (lsm != NULL), it will locate the stripe * by name hash (see lsm_name_to_stripe_info()). Note: if the hash_type * is unknown, it will return -EBADFD, and lmv_intent_lookup might need * walk through all of stripes to locate the entry. * * For normal direcotry, it will locate MDS by FID directly. * \param[in] lmv LMV device * \param[in] op_data client MD stack parameters, name, namelen * mds_num etc. * \param[in] fid object FID used to locate MDS. * * retval pointer to the lmv_tgt_desc if succeed. * ERR_PTR(errno) if failed. */ struct lmv_tgt_desc* lmv_locate_mds(struct lmv_obd *lmv, struct md_op_data *op_data, struct lu_fid *fid) { struct lmv_stripe_md *lsm = op_data->op_mea1; struct lmv_tgt_desc *tgt; /* During creating VOLATILE file, it should honor the mdt * index if the file under striped dir is being restored, see * ct_restore(). */ if (op_data->op_bias & MDS_CREATE_VOLATILE && (int)op_data->op_mds != -1) { int i; tgt = lmv_get_target(lmv, op_data->op_mds, NULL); if (IS_ERR(tgt)) return tgt; if (lsm != NULL) { /* refill the right parent fid */ for (i = 0; i < lsm->lsm_md_stripe_count; i++) { struct lmv_oinfo *oinfo; oinfo = &lsm->lsm_md_oinfo[i]; if (oinfo->lmo_mds == op_data->op_mds) { *fid = oinfo->lmo_fid; break; } } if (i == lsm->lsm_md_stripe_count) *fid = lsm->lsm_md_oinfo[0].lmo_fid; } return tgt; } if (lsm == NULL || op_data->op_namelen == 0) { tgt = lmv_find_target(lmv, fid); if (IS_ERR(tgt)) return tgt; op_data->op_mds = tgt->ltd_idx; return tgt; } return lmv_locate_target_for_name(lmv, lsm, op_data->op_name, op_data->op_namelen, fid, &op_data->op_mds); } int lmv_create(struct obd_export *exp, struct md_op_data *op_data, const void *data, size_t datalen, umode_t mode, uid_t uid, gid_t gid, cfs_cap_t cap_effective, __u64 rdev, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); if (!lmv->desc.ld_active_tgt_count) RETURN(-EIO); tgt = lmv_locate_mds(lmv, op_data, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); CDEBUG(D_INODE, "CREATE name '%.*s' on "DFID" -> mds #%x\n", (int)op_data->op_namelen, op_data->op_name, PFID(&op_data->op_fid1), op_data->op_mds); rc = lmv_fid_alloc(NULL, exp, &op_data->op_fid2, op_data); if (rc) RETURN(rc); if (exp_connect_flags(exp) & OBD_CONNECT_DIR_STRIPE) { /* Send the create request to the MDT where the object * will be located */ tgt = lmv_find_target(lmv, &op_data->op_fid2); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); op_data->op_mds = tgt->ltd_idx; } else { CDEBUG(D_CONFIG, "Server doesn't support striped dirs\n"); } CDEBUG(D_INODE, "CREATE obj "DFID" -> mds #%x\n", PFID(&op_data->op_fid2), op_data->op_mds); op_data->op_flags |= MF_MDC_CANCEL_FID1; rc = md_create(tgt->ltd_exp, op_data, data, datalen, mode, uid, gid, cap_effective, rdev, request); if (rc == 0) { if (*request == NULL) RETURN(rc); CDEBUG(D_INODE, "Created - "DFID"\n", PFID(&op_data->op_fid2)); } RETURN(rc); } static int lmv_enqueue(struct obd_export *exp, struct ldlm_enqueue_info *einfo, const union ldlm_policy_data *policy, struct lookup_intent *it, struct md_op_data *op_data, struct lustre_handle *lockh, __u64 extra_lock_flags) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); CDEBUG(D_INODE, "ENQUEUE '%s' on "DFID"\n", LL_IT2STR(it), PFID(&op_data->op_fid1)); tgt = lmv_locate_mds(lmv, op_data, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); CDEBUG(D_INODE, "ENQUEUE '%s' on "DFID" -> mds #%u\n", LL_IT2STR(it), PFID(&op_data->op_fid1), tgt->ltd_idx); rc = md_enqueue(tgt->ltd_exp, einfo, policy, it, op_data, lockh, extra_lock_flags); RETURN(rc); } static int lmv_getattr_name(struct obd_export *exp,struct md_op_data *op_data, struct ptlrpc_request **preq) { struct ptlrpc_request *req = NULL; struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; struct mdt_body *body; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); tgt = lmv_locate_mds(lmv, op_data, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); CDEBUG(D_INODE, "GETATTR_NAME for %*s on "DFID" -> mds #%d\n", (int)op_data->op_namelen, op_data->op_name, PFID(&op_data->op_fid1), tgt->ltd_idx); rc = md_getattr_name(tgt->ltd_exp, op_data, preq); if (rc != 0) RETURN(rc); body = req_capsule_server_get(&(*preq)->rq_pill, &RMF_MDT_BODY); LASSERT(body != NULL); if (body->mbo_valid & OBD_MD_MDS) { struct lu_fid rid = body->mbo_fid1; CDEBUG(D_INODE, "Request attrs for "DFID"\n", PFID(&rid)); tgt = lmv_find_target(lmv, &rid); if (IS_ERR(tgt)) { ptlrpc_req_finished(*preq); preq = NULL; RETURN(PTR_ERR(tgt)); } op_data->op_fid1 = rid; op_data->op_valid |= OBD_MD_FLCROSSREF; op_data->op_namelen = 0; op_data->op_name = NULL; rc = md_getattr_name(tgt->ltd_exp, op_data, &req); ptlrpc_req_finished(*preq); *preq = req; } RETURN(rc); } #define md_op_data_fid(op_data, fl) \ (fl == MF_MDC_CANCEL_FID1 ? &op_data->op_fid1 : \ fl == MF_MDC_CANCEL_FID2 ? &op_data->op_fid2 : \ fl == MF_MDC_CANCEL_FID3 ? &op_data->op_fid3 : \ fl == MF_MDC_CANCEL_FID4 ? &op_data->op_fid4 : \ NULL) static int lmv_early_cancel(struct obd_export *exp, struct lmv_tgt_desc *tgt, struct md_op_data *op_data, __u32 op_tgt, enum ldlm_mode mode, int bits, int flag) { struct lu_fid *fid = md_op_data_fid(op_data, flag); struct lmv_obd *lmv = &exp->exp_obd->u.lmv; union ldlm_policy_data policy = { { 0 } }; int rc = 0; ENTRY; if (!fid_is_sane(fid)) RETURN(0); if (tgt == NULL) { tgt = lmv_find_target(lmv, fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); } if (tgt->ltd_idx != op_tgt) { CDEBUG(D_INODE, "EARLY_CANCEL on "DFID"\n", PFID(fid)); policy.l_inodebits.bits = bits; rc = md_cancel_unused(tgt->ltd_exp, fid, &policy, mode, LCF_ASYNC, NULL); } else { CDEBUG(D_INODE, "EARLY_CANCEL skip operation target %d on "DFID"\n", op_tgt, PFID(fid)); op_data->op_flags |= flag; rc = 0; } RETURN(rc); } /* * llite passes fid of an target inode in op_data->op_fid1 and id of directory in * op_data->op_fid2 */ static int lmv_link(struct obd_export *exp, struct md_op_data *op_data, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); LASSERT(op_data->op_namelen != 0); CDEBUG(D_INODE, "LINK "DFID":%*s to "DFID"\n", PFID(&op_data->op_fid2), (int)op_data->op_namelen, op_data->op_name, PFID(&op_data->op_fid1)); op_data->op_fsuid = from_kuid(&init_user_ns, current_fsuid()); op_data->op_fsgid = from_kgid(&init_user_ns, current_fsgid()); op_data->op_cap = cfs_curproc_cap_pack(); if (op_data->op_mea2 != NULL) { struct lmv_stripe_md *lsm = op_data->op_mea2; const struct lmv_oinfo *oinfo; oinfo = lsm_name_to_stripe_info(lsm, op_data->op_name, op_data->op_namelen); if (IS_ERR(oinfo)) RETURN(PTR_ERR(oinfo)); op_data->op_fid2 = oinfo->lmo_fid; } tgt = lmv_locate_mds(lmv, op_data, &op_data->op_fid2); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); /* * Cancel UPDATE lock on child (fid1). */ op_data->op_flags |= MF_MDC_CANCEL_FID2; rc = lmv_early_cancel(exp, NULL, op_data, tgt->ltd_idx, LCK_EX, MDS_INODELOCK_UPDATE, MF_MDC_CANCEL_FID1); if (rc != 0) RETURN(rc); rc = md_link(tgt->ltd_exp, op_data, request); RETURN(rc); } static int lmv_rename(struct obd_export *exp, struct md_op_data *op_data, const char *old, size_t oldlen, const char *new, size_t newlen, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *src_tgt; struct lmv_tgt_desc *tgt_tgt; struct obd_export *target_exp; struct mdt_body *body; int rc; ENTRY; LASSERT(oldlen != 0); CDEBUG(D_INODE, "RENAME %.*s in "DFID":%d to %.*s in "DFID":%d\n", (int)oldlen, old, PFID(&op_data->op_fid1), op_data->op_mea1 ? op_data->op_mea1->lsm_md_stripe_count : 0, (int)newlen, new, PFID(&op_data->op_fid2), op_data->op_mea2 ? op_data->op_mea2->lsm_md_stripe_count : 0); rc = lmv_check_connect(obd); if (rc) RETURN(rc); op_data->op_fsuid = from_kuid(&init_user_ns, current_fsuid()); op_data->op_fsgid = from_kgid(&init_user_ns, current_fsgid()); op_data->op_cap = cfs_curproc_cap_pack(); if (op_data->op_cli_flags & CLI_MIGRATE) { LASSERTF(fid_is_sane(&op_data->op_fid3), "invalid FID "DFID"\n", PFID(&op_data->op_fid3)); if (op_data->op_mea1 != NULL) { struct lmv_stripe_md *lsm = op_data->op_mea1; struct lmv_tgt_desc *tmp; /* Fix the parent fid for striped dir */ tmp = lmv_locate_target_for_name(lmv, lsm, old, oldlen, &op_data->op_fid1, NULL); if (IS_ERR(tmp)) RETURN(PTR_ERR(tmp)); } rc = lmv_fid_alloc(NULL, exp, &op_data->op_fid2, op_data); if (rc != 0) RETURN(rc); src_tgt = lmv_find_target(lmv, &op_data->op_fid3); if (IS_ERR(src_tgt)) RETURN(PTR_ERR(src_tgt)); target_exp = src_tgt->ltd_exp; } else { if (op_data->op_mea1 != NULL) { struct lmv_stripe_md *lsm = op_data->op_mea1; src_tgt = lmv_locate_target_for_name(lmv, lsm, old, oldlen, &op_data->op_fid1, &op_data->op_mds); } else { src_tgt = lmv_find_target(lmv, &op_data->op_fid1); } if (IS_ERR(src_tgt)) RETURN(PTR_ERR(src_tgt)); if (op_data->op_mea2 != NULL) { struct lmv_stripe_md *lsm = op_data->op_mea2; tgt_tgt = lmv_locate_target_for_name(lmv, lsm, new, newlen, &op_data->op_fid2, &op_data->op_mds); } else { tgt_tgt = lmv_find_target(lmv, &op_data->op_fid2); } if (IS_ERR(tgt_tgt)) RETURN(PTR_ERR(tgt_tgt)); target_exp = tgt_tgt->ltd_exp; } /* * LOOKUP lock on src child (fid3) should also be cancelled for * src_tgt in mdc_rename. */ op_data->op_flags |= MF_MDC_CANCEL_FID1 | MF_MDC_CANCEL_FID3; /* * Cancel UPDATE locks on tgt parent (fid2), tgt_tgt is its * own target. */ rc = lmv_early_cancel(exp, NULL, op_data, src_tgt->ltd_idx, LCK_EX, MDS_INODELOCK_UPDATE, MF_MDC_CANCEL_FID2); if (rc != 0) RETURN(rc); /* * Cancel LOOKUP locks on source child (fid3) for parent tgt_tgt. */ if (fid_is_sane(&op_data->op_fid3)) { struct lmv_tgt_desc *tgt; tgt = lmv_find_target(lmv, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); /* Cancel LOOKUP lock on its parent */ rc = lmv_early_cancel(exp, tgt, op_data, src_tgt->ltd_idx, LCK_EX, MDS_INODELOCK_LOOKUP, MF_MDC_CANCEL_FID3); if (rc != 0) RETURN(rc); rc = lmv_early_cancel(exp, NULL, op_data, src_tgt->ltd_idx, LCK_EX, MDS_INODELOCK_FULL, MF_MDC_CANCEL_FID3); if (rc != 0) RETURN(rc); } retry_rename: /* * Cancel all the locks on tgt child (fid4). */ if (fid_is_sane(&op_data->op_fid4)) { struct lmv_tgt_desc *tgt; rc = lmv_early_cancel(exp, NULL, op_data, src_tgt->ltd_idx, LCK_EX, MDS_INODELOCK_FULL, MF_MDC_CANCEL_FID4); if (rc != 0) RETURN(rc); tgt = lmv_find_target(lmv, &op_data->op_fid4); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); /* Since the target child might be destroyed, and it might * become orphan, and we can only check orphan on the local * MDT right now, so we send rename request to the MDT where * target child is located. If target child does not exist, * then it will send the request to the target parent */ target_exp = tgt->ltd_exp; } rc = md_rename(target_exp, op_data, old, oldlen, new, newlen, request); if (rc != 0 && rc != -EXDEV) RETURN(rc); body = req_capsule_server_get(&(*request)->rq_pill, &RMF_MDT_BODY); if (body == NULL) RETURN(-EPROTO); /* Not cross-ref case, just get out of here. */ if (likely(!(body->mbo_valid & OBD_MD_MDS))) RETURN(rc); CDEBUG(D_INODE, "%s: try rename to another MDT for "DFID"\n", exp->exp_obd->obd_name, PFID(&body->mbo_fid1)); op_data->op_fid4 = body->mbo_fid1; ptlrpc_req_finished(*request); *request = NULL; goto retry_rename; } static int lmv_setattr(struct obd_export *exp, struct md_op_data *op_data, void *ea, size_t ealen, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc = 0; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); CDEBUG(D_INODE, "SETATTR for "DFID", valid 0x%x\n", PFID(&op_data->op_fid1), op_data->op_attr.ia_valid); op_data->op_flags |= MF_MDC_CANCEL_FID1; tgt = lmv_find_target(lmv, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); rc = md_setattr(tgt->ltd_exp, op_data, ea, ealen, request); RETURN(rc); } static int lmv_fsync(struct obd_export *exp, const struct lu_fid *fid, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc != 0) RETURN(rc); tgt = lmv_find_target(lmv, fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); rc = md_fsync(tgt->ltd_exp, fid, request); RETURN(rc); } /** * Get current minimum entry from striped directory * * This function will search the dir entry, whose hash value is the * closest(>=) to @hash_offset, from all of sub-stripes, and it is * only being called for striped directory. * * \param[in] exp export of LMV * \param[in] op_data parameters transferred beween client MD stack * stripe_information will be included in this * parameter * \param[in] cb_op ldlm callback being used in enqueue in * mdc_read_page * \param[in] hash_offset the hash value, which is used to locate * minum(closet) dir entry * \param[in|out] stripe_offset the caller use this to indicate the stripe * index of last entry, so to avoid hash conflict * between stripes. It will also be used to * return the stripe index of current dir entry. * \param[in|out] entp the minum entry and it also is being used * to input the last dir entry to resolve the * hash conflict * * \param[out] ppage the page which holds the minum entry * * \retval = 0 get the entry successfully * negative errno (< 0) does not get the entry */ static int lmv_get_min_striped_entry(struct obd_export *exp, struct md_op_data *op_data, struct md_callback *cb_op, __u64 hash_offset, int *stripe_offset, struct lu_dirent **entp, struct page **ppage) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_stripe_md *lsm = op_data->op_mea1; struct lmv_tgt_desc *tgt; int stripe_count; struct lu_dirent *min_ent = NULL; struct page *min_page = NULL; int min_idx = 0; int i; int rc = 0; ENTRY; stripe_count = lsm->lsm_md_stripe_count; for (i = 0; i < stripe_count; i++) { struct lu_dirent *ent = NULL; struct page *page = NULL; struct lu_dirpage *dp; __u64 stripe_hash = hash_offset; tgt = lmv_get_target(lmv, lsm->lsm_md_oinfo[i].lmo_mds, NULL); if (IS_ERR(tgt)) GOTO(out, rc = PTR_ERR(tgt)); /* op_data will be shared by each stripe, so we need * reset these value for each stripe */ op_data->op_fid1 = lsm->lsm_md_oinfo[i].lmo_fid; op_data->op_fid2 = lsm->lsm_md_oinfo[i].lmo_fid; op_data->op_data = lsm->lsm_md_oinfo[i].lmo_root; next: rc = md_read_page(tgt->ltd_exp, op_data, cb_op, stripe_hash, &page); if (rc != 0) GOTO(out, rc); dp = page_address(page); for (ent = lu_dirent_start(dp); ent != NULL; ent = lu_dirent_next(ent)) { /* Skip dummy entry */ if (le16_to_cpu(ent->lde_namelen) == 0) continue; if (le64_to_cpu(ent->lde_hash) < hash_offset) continue; if (le64_to_cpu(ent->lde_hash) == hash_offset && (*entp == ent || i < *stripe_offset)) continue; /* skip . and .. for other stripes */ if (i != 0 && (strncmp(ent->lde_name, ".", le16_to_cpu(ent->lde_namelen)) == 0 || strncmp(ent->lde_name, "..", le16_to_cpu(ent->lde_namelen)) == 0)) continue; break; } if (ent == NULL) { stripe_hash = le64_to_cpu(dp->ldp_hash_end); kunmap(page); page_cache_release(page); page = NULL; /* reach the end of current stripe, go to next stripe */ if (stripe_hash == MDS_DIR_END_OFF) continue; else goto next; } if (min_ent != NULL) { if (le64_to_cpu(min_ent->lde_hash) > le64_to_cpu(ent->lde_hash)) { min_ent = ent; kunmap(min_page); page_cache_release(min_page); min_idx = i; min_page = page; } else { kunmap(page); page_cache_release(page); page = NULL; } } else { min_ent = ent; min_page = page; min_idx = i; } } out: if (*ppage != NULL) { kunmap(*ppage); page_cache_release(*ppage); } *stripe_offset = min_idx; *entp = min_ent; *ppage = min_page; RETURN(rc); } /** * Build dir entry page from a striped directory * * This function gets one entry by @offset from a striped directory. It will * read entries from all of stripes, and choose one closest to the required * offset(&offset). A few notes * 1. skip . and .. for non-zero stripes, because there can only have one . * and .. in a directory. * 2. op_data will be shared by all of stripes, instead of allocating new * one, so need to restore before reusing. * 3. release the entry page if that is not being chosen. * * \param[in] exp obd export refer to LMV * \param[in] op_data hold those MD parameters of read_entry * \param[in] cb_op ldlm callback being used in enqueue in mdc_read_entry * \param[out] ldp the entry being read * \param[out] ppage the page holding the entry. Note: because the entry * will be accessed in upper layer, so we need hold the * page until the usages of entry is finished, see * ll_dir_entry_next. * * retval =0 if get entry successfully * <0 cannot get entry */ static int lmv_read_striped_page(struct obd_export *exp, struct md_op_data *op_data, struct md_callback *cb_op, __u64 offset, struct page **ppage) { struct obd_device *obd = exp->exp_obd; struct lu_fid master_fid = op_data->op_fid1; struct inode *master_inode = op_data->op_data; __u64 hash_offset = offset; struct lu_dirpage *dp; struct page *min_ent_page = NULL; struct page *ent_page = NULL; struct lu_dirent *ent; void *area; int ent_idx = 0; struct lu_dirent *min_ent = NULL; struct lu_dirent *last_ent; size_t left_bytes; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); /* Allocate a page and read entries from all of stripes and fill * the page by hash order */ ent_page = alloc_page(GFP_KERNEL); if (ent_page == NULL) RETURN(-ENOMEM); /* Initialize the entry page */ dp = kmap(ent_page); memset(dp, 0, sizeof(*dp)); dp->ldp_hash_start = cpu_to_le64(offset); dp->ldp_flags |= LDF_COLLIDE; area = dp + 1; left_bytes = PAGE_CACHE_SIZE - sizeof(*dp); ent = area; last_ent = ent; do { __u16 ent_size; /* Find the minum entry from all sub-stripes */ rc = lmv_get_min_striped_entry(exp, op_data, cb_op, hash_offset, &ent_idx, &min_ent, &min_ent_page); if (rc != 0) GOTO(out, rc); /* If it can not get minum entry, it means it already reaches * the end of this directory */ if (min_ent == NULL) { last_ent->lde_reclen = 0; hash_offset = MDS_DIR_END_OFF; GOTO(out, rc); } ent_size = le16_to_cpu(min_ent->lde_reclen); /* the last entry lde_reclen is 0, but it might not * the end of this entry of this temporay entry */ if (ent_size == 0) ent_size = lu_dirent_calc_size( le16_to_cpu(min_ent->lde_namelen), le32_to_cpu(min_ent->lde_attrs)); if (ent_size > left_bytes) { last_ent->lde_reclen = cpu_to_le16(0); hash_offset = le64_to_cpu(min_ent->lde_hash); GOTO(out, rc); } memcpy(ent, min_ent, ent_size); /* Replace . with master FID and Replace .. with the parent FID * of master object */ if (strncmp(ent->lde_name, ".", le16_to_cpu(ent->lde_namelen)) == 0 && le16_to_cpu(ent->lde_namelen) == 1) fid_cpu_to_le(&ent->lde_fid, &master_fid); else if (strncmp(ent->lde_name, "..", le16_to_cpu(ent->lde_namelen)) == 0 && le16_to_cpu(ent->lde_namelen) == 2) fid_cpu_to_le(&ent->lde_fid, &op_data->op_fid3); left_bytes -= ent_size; ent->lde_reclen = cpu_to_le16(ent_size); last_ent = ent; ent = (void *)ent + ent_size; hash_offset = le64_to_cpu(min_ent->lde_hash); if (hash_offset == MDS_DIR_END_OFF) { last_ent->lde_reclen = 0; break; } } while (1); out: if (min_ent_page != NULL) { kunmap(min_ent_page); page_cache_release(min_ent_page); } if (unlikely(rc != 0)) { __free_page(ent_page); ent_page = NULL; } else { if (ent == area) dp->ldp_flags |= LDF_EMPTY; dp->ldp_flags = cpu_to_le32(dp->ldp_flags); dp->ldp_hash_end = cpu_to_le64(hash_offset); } /* We do not want to allocate md_op_data during each * dir entry reading, so op_data will be shared by every stripe, * then we need to restore it back to original value before * return to the upper layer */ op_data->op_fid1 = master_fid; op_data->op_fid2 = master_fid; op_data->op_data = master_inode; *ppage = ent_page; RETURN(rc); } int lmv_read_page(struct obd_export *exp, struct md_op_data *op_data, struct md_callback *cb_op, __u64 offset, struct page **ppage) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_stripe_md *lsm = op_data->op_mea1; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc != 0) RETURN(rc); if (unlikely(lsm != NULL)) { rc = lmv_read_striped_page(exp, op_data, cb_op, offset, ppage); RETURN(rc); } tgt = lmv_find_target(lmv, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); rc = md_read_page(tgt->ltd_exp, op_data, cb_op, offset, ppage); RETURN(rc); } /** * Unlink a file/directory * * Unlink a file or directory under the parent dir. The unlink request * usually will be sent to the MDT where the child is located, but if * the client does not have the child FID then request will be sent to the * MDT where the parent is located. * * If the parent is a striped directory then it also needs to locate which * stripe the name of the child is located, and replace the parent FID * (@op->op_fid1) with the stripe FID. Note: if the stripe is unknown, * it will walk through all of sub-stripes until the child is being * unlinked finally. * * \param[in] exp export refer to LMV * \param[in] op_data different parameters transferred beween client * MD stacks, name, namelen, FIDs etc. * op_fid1 is the parent FID, op_fid2 is the child * FID. * \param[out] request point to the request of unlink. * * retval 0 if succeed * negative errno if failed. */ static int lmv_unlink(struct obd_export *exp, struct md_op_data *op_data, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt = NULL; struct lmv_tgt_desc *parent_tgt = NULL; struct mdt_body *body; int rc; int stripe_index = 0; struct lmv_stripe_md *lsm = op_data->op_mea1; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); retry_unlink: /* For striped dir, we need to locate the parent as well */ if (lsm != NULL) { struct lmv_tgt_desc *tmp; LASSERT(op_data->op_name != NULL && op_data->op_namelen != 0); tmp = lmv_locate_target_for_name(lmv, lsm, op_data->op_name, op_data->op_namelen, &op_data->op_fid1, &op_data->op_mds); /* return -EBADFD means unknown hash type, might * need try all sub-stripe here */ if (IS_ERR(tmp) && PTR_ERR(tmp) != -EBADFD) RETURN(PTR_ERR(tmp)); /* Note: both migrating dir and unknown hash dir need to * try all of sub-stripes, so we need start search the * name from stripe 0, but migrating dir is already handled * inside lmv_locate_target_for_name(), so we only check * unknown hash type directory here */ if (!lmv_is_known_hash_type(lsm->lsm_md_hash_type)) { struct lmv_oinfo *oinfo; oinfo = &lsm->lsm_md_oinfo[stripe_index]; op_data->op_fid1 = oinfo->lmo_fid; op_data->op_mds = oinfo->lmo_mds; } } try_next_stripe: /* Send unlink requests to the MDT where the child is located */ if (likely(!fid_is_zero(&op_data->op_fid2))) tgt = lmv_find_target(lmv, &op_data->op_fid2); else if (lsm != NULL) tgt = lmv_get_target(lmv, op_data->op_mds, NULL); else tgt = lmv_locate_mds(lmv, op_data, &op_data->op_fid1); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); op_data->op_fsuid = from_kuid(&init_user_ns, current_fsuid()); op_data->op_fsgid = from_kgid(&init_user_ns, current_fsgid()); op_data->op_cap = cfs_curproc_cap_pack(); /* * If child's fid is given, cancel unused locks for it if it is from * another export than parent. * * LOOKUP lock for child (fid3) should also be cancelled on parent * tgt_tgt in mdc_unlink(). */ op_data->op_flags |= MF_MDC_CANCEL_FID1 | MF_MDC_CANCEL_FID3; /* * Cancel FULL locks on child (fid3). */ parent_tgt = lmv_find_target(lmv, &op_data->op_fid1); if (IS_ERR(parent_tgt)) RETURN(PTR_ERR(parent_tgt)); if (parent_tgt != tgt) { rc = lmv_early_cancel(exp, parent_tgt, op_data, tgt->ltd_idx, LCK_EX, MDS_INODELOCK_LOOKUP, MF_MDC_CANCEL_FID3); } rc = lmv_early_cancel(exp, NULL, op_data, tgt->ltd_idx, LCK_EX, MDS_INODELOCK_FULL, MF_MDC_CANCEL_FID3); if (rc != 0) RETURN(rc); CDEBUG(D_INODE, "unlink with fid="DFID"/"DFID" -> mds #%u\n", PFID(&op_data->op_fid1), PFID(&op_data->op_fid2), tgt->ltd_idx); rc = md_unlink(tgt->ltd_exp, op_data, request); if (rc != 0 && rc != -EREMOTE && rc != -ENOENT) RETURN(rc); /* Try next stripe if it is needed. */ if (rc == -ENOENT && lsm != NULL && lmv_need_try_all_stripes(lsm)) { struct lmv_oinfo *oinfo; stripe_index++; if (stripe_index >= lsm->lsm_md_stripe_count) RETURN(rc); oinfo = &lsm->lsm_md_oinfo[stripe_index]; op_data->op_fid1 = oinfo->lmo_fid; op_data->op_mds = oinfo->lmo_mds; ptlrpc_req_finished(*request); *request = NULL; goto try_next_stripe; } body = req_capsule_server_get(&(*request)->rq_pill, &RMF_MDT_BODY); if (body == NULL) RETURN(-EPROTO); /* Not cross-ref case, just get out of here. */ if (likely(!(body->mbo_valid & OBD_MD_MDS))) RETURN(rc); CDEBUG(D_INODE, "%s: try unlink to another MDT for "DFID"\n", exp->exp_obd->obd_name, PFID(&body->mbo_fid1)); /* This is a remote object, try remote MDT, Note: it may * try more than 1 time here, Considering following case * /mnt/lustre is root on MDT0, remote1 is on MDT1 * 1. Initially A does not know where remote1 is, it send * unlink RPC to MDT0, MDT0 return -EREMOTE, it will * resend unlink RPC to MDT1 (retry 1st time). * * 2. During the unlink RPC in flight, * client B mv /mnt/lustre/remote1 /mnt/lustre/remote2 * and create new remote1, but on MDT0 * * 3. MDT1 get unlink RPC(from A), then do remote lock on * /mnt/lustre, then lookup get fid of remote1, and find * it is remote dir again, and replay -EREMOTE again. * * 4. Then A will resend unlink RPC to MDT0. (retry 2nd times). * * In theory, it might try unlimited time here, but it should * be very rare case. */ op_data->op_fid2 = body->mbo_fid1; ptlrpc_req_finished(*request); *request = NULL; goto retry_unlink; } static int lmv_precleanup(struct obd_device *obd) { ENTRY; fld_client_proc_fini(&obd->u.lmv.lmv_fld); lprocfs_obd_cleanup(obd); lprocfs_free_md_stats(obd); RETURN(0); } /** * Get by key a value associated with a LMV device. * * Dispatch request to lower-layer devices as needed. * * \param[in] env execution environment for this thread * \param[in] exp export for the LMV device * \param[in] keylen length of key identifier * \param[in] key identifier of key to get value for * \param[in] vallen size of \a val * \param[out] val pointer to storage location for value * \param[in] lsm optional striping metadata of object * * \retval 0 on success * \retval negative negated errno on failure */ static int lmv_get_info(const struct lu_env *env, struct obd_export *exp, __u32 keylen, void *key, __u32 *vallen, void *val) { struct obd_device *obd; struct lmv_obd *lmv; int rc = 0; ENTRY; obd = class_exp2obd(exp); if (obd == NULL) { CDEBUG(D_IOCTL, "Invalid client cookie "LPX64"\n", exp->exp_handle.h_cookie); RETURN(-EINVAL); } lmv = &obd->u.lmv; if (keylen >= strlen("remote_flag") && !strcmp(key, "remote_flag")) { int i; rc = lmv_check_connect(obd); if (rc) RETURN(rc); LASSERT(*vallen == sizeof(__u32)); for (i = 0; i < lmv->desc.ld_tgt_count; i++) { struct lmv_tgt_desc *tgt = lmv->tgts[i]; /* * All tgts should be connected when this gets called. */ if (tgt == NULL || tgt->ltd_exp == NULL) continue; if (!obd_get_info(env, tgt->ltd_exp, keylen, key, vallen, val)) RETURN(0); } RETURN(-EINVAL); } else if (KEY_IS(KEY_MAX_EASIZE) || KEY_IS(KEY_DEFAULT_EASIZE) || KEY_IS(KEY_CONN_DATA)) { rc = lmv_check_connect(obd); if (rc) RETURN(rc); /* * Forwarding this request to first MDS, it should know LOV * desc. */ rc = obd_get_info(env, lmv->tgts[0]->ltd_exp, keylen, key, vallen, val); if (!rc && KEY_IS(KEY_CONN_DATA)) exp->exp_connect_data = *(struct obd_connect_data *)val; RETURN(rc); } else if (KEY_IS(KEY_TGT_COUNT)) { *((int *)val) = lmv->desc.ld_tgt_count; RETURN(0); } CDEBUG(D_IOCTL, "Invalid key\n"); RETURN(-EINVAL); } /** * Asynchronously set by key a value associated with a LMV device. * * Dispatch request to lower-layer devices as needed. * * \param[in] env execution environment for this thread * \param[in] exp export for the LMV device * \param[in] keylen length of key identifier * \param[in] key identifier of key to store value for * \param[in] vallen size of value to store * \param[in] val pointer to data to be stored * \param[in] set optional list of related ptlrpc requests * * \retval 0 on success * \retval negative negated errno on failure */ int lmv_set_info_async(const struct lu_env *env, struct obd_export *exp, __u32 keylen, void *key, __u32 vallen, void *val, struct ptlrpc_request_set *set) { struct lmv_tgt_desc *tgt = NULL; struct obd_device *obd; struct lmv_obd *lmv; int rc = 0; ENTRY; obd = class_exp2obd(exp); if (obd == NULL) { CDEBUG(D_IOCTL, "Invalid client cookie "LPX64"\n", exp->exp_handle.h_cookie); RETURN(-EINVAL); } lmv = &obd->u.lmv; if (KEY_IS(KEY_READ_ONLY) || KEY_IS(KEY_FLUSH_CTX) || KEY_IS(KEY_DEFAULT_EASIZE)) { int i, err = 0; for (i = 0; i < lmv->desc.ld_tgt_count; i++) { tgt = lmv->tgts[i]; if (tgt == NULL || tgt->ltd_exp == NULL) continue; err = obd_set_info_async(env, tgt->ltd_exp, keylen, key, vallen, val, set); if (err && rc == 0) rc = err; } RETURN(rc); } RETURN(-EINVAL); } static int lmv_unpack_md_v1(struct obd_export *exp, struct lmv_stripe_md *lsm, const struct lmv_mds_md_v1 *lmm1) { struct lmv_obd *lmv = &exp->exp_obd->u.lmv; int stripe_count; int cplen; int i; int rc = 0; ENTRY; lsm->lsm_md_magic = le32_to_cpu(lmm1->lmv_magic); lsm->lsm_md_stripe_count = le32_to_cpu(lmm1->lmv_stripe_count); lsm->lsm_md_master_mdt_index = le32_to_cpu(lmm1->lmv_master_mdt_index); if (OBD_FAIL_CHECK(OBD_FAIL_UNKNOWN_LMV_STRIPE)) lsm->lsm_md_hash_type = LMV_HASH_TYPE_UNKNOWN; else lsm->lsm_md_hash_type = le32_to_cpu(lmm1->lmv_hash_type); lsm->lsm_md_layout_version = le32_to_cpu(lmm1->lmv_layout_version); cplen = strlcpy(lsm->lsm_md_pool_name, lmm1->lmv_pool_name, sizeof(lsm->lsm_md_pool_name)); if (cplen >= sizeof(lsm->lsm_md_pool_name)) RETURN(-E2BIG); CDEBUG(D_INFO, "unpack lsm count %d, master %d hash_type %d" "layout_version %d\n", lsm->lsm_md_stripe_count, lsm->lsm_md_master_mdt_index, lsm->lsm_md_hash_type, lsm->lsm_md_layout_version); stripe_count = le32_to_cpu(lmm1->lmv_stripe_count); for (i = 0; i < stripe_count; i++) { fid_le_to_cpu(&lsm->lsm_md_oinfo[i].lmo_fid, &lmm1->lmv_stripe_fids[i]); rc = lmv_fld_lookup(lmv, &lsm->lsm_md_oinfo[i].lmo_fid, &lsm->lsm_md_oinfo[i].lmo_mds); if (rc != 0) RETURN(rc); CDEBUG(D_INFO, "unpack fid #%d "DFID"\n", i, PFID(&lsm->lsm_md_oinfo[i].lmo_fid)); } RETURN(rc); } static int lmv_unpackmd(struct obd_export *exp, struct lmv_stripe_md **lsmp, const union lmv_mds_md *lmm, size_t lmm_size) { struct lmv_stripe_md *lsm; int lsm_size; int rc; bool allocated = false; ENTRY; LASSERT(lsmp != NULL); lsm = *lsmp; /* Free memmd */ if (lsm != NULL && lmm == NULL) { int i; for (i = 0; i < lsm->lsm_md_stripe_count; i++) { /* For migrating inode, the master stripe and master * object will be the same, so do not need iput, see * ll_update_lsm_md */ if (!(lsm->lsm_md_hash_type & LMV_HASH_FLAG_MIGRATION && i == 0) && lsm->lsm_md_oinfo[i].lmo_root != NULL) iput(lsm->lsm_md_oinfo[i].lmo_root); } lsm_size = lmv_stripe_md_size(lsm->lsm_md_stripe_count); OBD_FREE(lsm, lsm_size); *lsmp = NULL; RETURN(0); } if (le32_to_cpu(lmm->lmv_magic) == LMV_MAGIC_STRIPE) RETURN(-EPERM); /* Unpack memmd */ if (le32_to_cpu(lmm->lmv_magic) != LMV_MAGIC_V1 && le32_to_cpu(lmm->lmv_magic) != LMV_USER_MAGIC) { CERROR("%s: invalid lmv magic %x: rc = %d\n", exp->exp_obd->obd_name, le32_to_cpu(lmm->lmv_magic), -EIO); RETURN(-EIO); } if (le32_to_cpu(lmm->lmv_magic) == LMV_MAGIC_V1) lsm_size = lmv_stripe_md_size(lmv_mds_md_stripe_count_get(lmm)); else /** * Unpack default dirstripe(lmv_user_md) to lmv_stripe_md, * stripecount should be 0 then. */ lsm_size = lmv_stripe_md_size(0); lsm_size = lmv_stripe_md_size(lmv_mds_md_stripe_count_get(lmm)); if (lsm == NULL) { OBD_ALLOC(lsm, lsm_size); if (lsm == NULL) RETURN(-ENOMEM); allocated = true; *lsmp = lsm; } switch (le32_to_cpu(lmm->lmv_magic)) { case LMV_MAGIC_V1: rc = lmv_unpack_md_v1(exp, lsm, &lmm->lmv_md_v1); break; default: CERROR("%s: unrecognized magic %x\n", exp->exp_obd->obd_name, le32_to_cpu(lmm->lmv_magic)); rc = -EINVAL; break; } if (rc != 0 && allocated) { OBD_FREE(lsm, lsm_size); *lsmp = NULL; lsm_size = rc; } RETURN(lsm_size); } void lmv_free_memmd(struct lmv_stripe_md *lsm) { lmv_unpackmd(NULL, &lsm, NULL, 0); } EXPORT_SYMBOL(lmv_free_memmd); static int lmv_cancel_unused(struct obd_export *exp, const struct lu_fid *fid, union ldlm_policy_data *policy, enum ldlm_mode mode, enum ldlm_cancel_flags flags, void *opaque) { struct lmv_obd *lmv = &exp->exp_obd->u.lmv; int rc = 0; __u32 i; ENTRY; LASSERT(fid != NULL); for (i = 0; i < lmv->desc.ld_tgt_count; i++) { struct lmv_tgt_desc *tgt = lmv->tgts[i]; int err; if (tgt == NULL || tgt->ltd_exp == NULL || !tgt->ltd_active) continue; err = md_cancel_unused(tgt->ltd_exp, fid, policy, mode, flags, opaque); if (!rc) rc = err; } RETURN(rc); } int lmv_set_lock_data(struct obd_export *exp, __u64 *lockh, void *data, __u64 *bits) { struct lmv_obd *lmv = &exp->exp_obd->u.lmv; struct lmv_tgt_desc *tgt = lmv->tgts[0]; int rc; ENTRY; if (tgt == NULL || tgt->ltd_exp == NULL) RETURN(-EINVAL); rc = md_set_lock_data(tgt->ltd_exp, lockh, data, bits); RETURN(rc); } enum ldlm_mode lmv_lock_match(struct obd_export *exp, __u64 flags, const struct lu_fid *fid, enum ldlm_type type, union ldlm_policy_data *policy, enum ldlm_mode mode, struct lustre_handle *lockh) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; enum ldlm_mode rc; int tgt; int i; ENTRY; CDEBUG(D_INODE, "Lock match for "DFID"\n", PFID(fid)); /* * With DNE every object can have two locks in different namespaces: * lookup lock in space of MDT storing direntry and update/open lock in * space of MDT storing inode. Try the MDT that the FID maps to first, * since this can be easily found, and only try others if that fails. */ for (i = 0, tgt = lmv_find_target_index(lmv, fid); i < lmv->desc.ld_tgt_count; i++, tgt = (tgt + 1) % lmv->desc.ld_tgt_count) { if (tgt < 0) { CDEBUG(D_HA, "%s: "DFID" is inaccessible: rc = %d\n", obd->obd_name, PFID(fid), tgt); tgt = 0; } if (lmv->tgts[tgt] == NULL || lmv->tgts[tgt]->ltd_exp == NULL || lmv->tgts[tgt]->ltd_active == 0) continue; rc = md_lock_match(lmv->tgts[tgt]->ltd_exp, flags, fid, type, policy, mode, lockh); if (rc) RETURN(rc); } RETURN(0); } int lmv_get_lustre_md(struct obd_export *exp, struct ptlrpc_request *req, struct obd_export *dt_exp, struct obd_export *md_exp, struct lustre_md *md) { struct lmv_obd *lmv = &exp->exp_obd->u.lmv; struct lmv_tgt_desc *tgt = lmv->tgts[0]; if (tgt == NULL || tgt->ltd_exp == NULL) RETURN(-EINVAL); return md_get_lustre_md(lmv->tgts[0]->ltd_exp, req, dt_exp, md_exp, md); } int lmv_free_lustre_md(struct obd_export *exp, struct lustre_md *md) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt = lmv->tgts[0]; ENTRY; if (md->lmv != NULL) { lmv_free_memmd(md->lmv); md->lmv = NULL; } if (tgt == NULL || tgt->ltd_exp == NULL) RETURN(-EINVAL); RETURN(md_free_lustre_md(lmv->tgts[0]->ltd_exp, md)); } int lmv_set_open_replay_data(struct obd_export *exp, struct obd_client_handle *och, struct lookup_intent *it) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; ENTRY; tgt = lmv_find_target(lmv, &och->och_fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); RETURN(md_set_open_replay_data(tgt->ltd_exp, och, it)); } int lmv_clear_open_replay_data(struct obd_export *exp, struct obd_client_handle *och) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; ENTRY; tgt = lmv_find_target(lmv, &och->och_fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); RETURN(md_clear_open_replay_data(tgt->ltd_exp, och)); } static int lmv_get_remote_perm(struct obd_export *exp, const struct lu_fid *fid, u32 suppgid, struct ptlrpc_request **request) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); tgt = lmv_find_target(lmv, fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); rc = md_get_remote_perm(tgt->ltd_exp, fid, suppgid, request); RETURN(rc); } int lmv_intent_getattr_async(struct obd_export *exp, struct md_enqueue_info *minfo) { struct md_op_data *op_data = &minfo->mi_data; struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *ptgt = NULL; struct lmv_tgt_desc *ctgt = NULL; int rc; ENTRY; if (!fid_is_sane(&op_data->op_fid2)) RETURN(-EINVAL); rc = lmv_check_connect(obd); if (rc) RETURN(rc); ptgt = lmv_locate_mds(lmv, op_data, &op_data->op_fid1); if (IS_ERR(ptgt)) RETURN(PTR_ERR(ptgt)); ctgt = lmv_locate_mds(lmv, op_data, &op_data->op_fid2); if (IS_ERR(ctgt)) RETURN(PTR_ERR(ctgt)); /* * if child is on remote MDT, we need 2 async RPCs to fetch both LOOKUP * lock on parent, and UPDATE lock on child MDT, which makes all * complicated. Considering remote dir is rare case, and not supporting * it in statahead won't cause any issue, drop its support for now. */ if (ptgt != ctgt) RETURN(-ENOTSUPP); rc = md_intent_getattr_async(ptgt->ltd_exp, minfo); RETURN(rc); } int lmv_revalidate_lock(struct obd_export *exp, struct lookup_intent *it, struct lu_fid *fid, __u64 *bits) { struct obd_device *obd = exp->exp_obd; struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt; int rc; ENTRY; rc = lmv_check_connect(obd); if (rc) RETURN(rc); tgt = lmv_find_target(lmv, fid); if (IS_ERR(tgt)) RETURN(PTR_ERR(tgt)); rc = md_revalidate_lock(tgt->ltd_exp, it, fid, bits); RETURN(rc); } int lmv_get_fid_from_lsm(struct obd_export *exp, const struct lmv_stripe_md *lsm, const char *name, int namelen, struct lu_fid *fid) { const struct lmv_oinfo *oinfo; LASSERT(lsm != NULL); oinfo = lsm_name_to_stripe_info(lsm, name, namelen); if (IS_ERR(oinfo)) return PTR_ERR(oinfo); *fid = oinfo->lmo_fid; RETURN(0); } /** * For lmv, only need to send request to master MDT, and the master MDT will * process with other slave MDTs. The only exception is Q_GETOQUOTA for which * we directly fetch data from the slave MDTs. */ int lmv_quotactl(struct obd_device *unused, struct obd_export *exp, struct obd_quotactl *oqctl) { struct obd_device *obd = class_exp2obd(exp); struct lmv_obd *lmv = &obd->u.lmv; struct lmv_tgt_desc *tgt = lmv->tgts[0]; int rc = 0; __u32 i; __u64 curspace, curinodes; ENTRY; if (tgt == NULL || tgt->ltd_exp == NULL || !tgt->ltd_active || lmv->desc.ld_tgt_count == 0) { CERROR("master lmv inactive\n"); RETURN(-EIO); } if (oqctl->qc_cmd != Q_GETOQUOTA) { rc = obd_quotactl(tgt->ltd_exp, oqctl); RETURN(rc); } curspace = curinodes = 0; for (i = 0; i < lmv->desc.ld_tgt_count; i++) { int err; tgt = lmv->tgts[i]; if (tgt == NULL || tgt->ltd_exp == NULL || !tgt->ltd_active) continue; err = obd_quotactl(tgt->ltd_exp, oqctl); if (err) { CERROR("getquota on mdt %d failed. %d\n", i, err); if (!rc) rc = err; } else { curspace += oqctl->qc_dqblk.dqb_curspace; curinodes += oqctl->qc_dqblk.dqb_curinodes; } } oqctl->qc_dqblk.dqb_curspace = curspace; oqctl->qc_dqblk.dqb_curinodes = curinodes; RETURN(rc); } static int lmv_merge_attr(struct obd_export *exp, const struct lmv_stripe_md *lsm, struct cl_attr *attr, ldlm_blocking_callback cb_blocking) { int rc; int i; rc = lmv_revalidate_slaves(exp, lsm, cb_blocking, 0); if (rc < 0) return rc; for (i = 0; i < lsm->lsm_md_stripe_count; i++) { struct inode *inode = lsm->lsm_md_oinfo[i].lmo_root; CDEBUG(D_INFO, ""DFID" size %llu, blocks %llu nlink %u," " atime %lu ctime %lu, mtime %lu.\n", PFID(&lsm->lsm_md_oinfo[i].lmo_fid), i_size_read(inode), (unsigned long long)inode->i_blocks, inode->i_nlink, LTIME_S(inode->i_atime), LTIME_S(inode->i_ctime), LTIME_S(inode->i_mtime)); /* for slave stripe, it needs to subtract nlink for . and .. */ if (i != 0) attr->cat_nlink += inode->i_nlink - 2; else attr->cat_nlink = inode->i_nlink; attr->cat_size += i_size_read(inode); attr->cat_blocks += inode->i_blocks; if (attr->cat_atime < LTIME_S(inode->i_atime)) attr->cat_atime = LTIME_S(inode->i_atime); if (attr->cat_ctime < LTIME_S(inode->i_ctime)) attr->cat_ctime = LTIME_S(inode->i_ctime); if (attr->cat_mtime < LTIME_S(inode->i_mtime)) attr->cat_mtime = LTIME_S(inode->i_mtime); } return 0; } struct obd_ops lmv_obd_ops = { .o_owner = THIS_MODULE, .o_setup = lmv_setup, .o_cleanup = lmv_cleanup, .o_precleanup = lmv_precleanup, .o_process_config = lmv_process_config, .o_connect = lmv_connect, .o_disconnect = lmv_disconnect, .o_statfs = lmv_statfs, .o_get_info = lmv_get_info, .o_set_info_async = lmv_set_info_async, .o_notify = lmv_notify, .o_get_uuid = lmv_get_uuid, .o_iocontrol = lmv_iocontrol, .o_quotactl = lmv_quotactl }; struct md_ops lmv_md_ops = { .m_getstatus = lmv_getstatus, .m_null_inode = lmv_null_inode, .m_find_cbdata = lmv_find_cbdata, .m_close = lmv_close, .m_create = lmv_create, .m_enqueue = lmv_enqueue, .m_getattr = lmv_getattr, .m_getxattr = lmv_getxattr, .m_getattr_name = lmv_getattr_name, .m_intent_lock = lmv_intent_lock, .m_link = lmv_link, .m_rename = lmv_rename, .m_setattr = lmv_setattr, .m_setxattr = lmv_setxattr, .m_fsync = lmv_fsync, .m_read_page = lmv_read_page, .m_unlink = lmv_unlink, .m_init_ea_size = lmv_init_ea_size, .m_cancel_unused = lmv_cancel_unused, .m_set_lock_data = lmv_set_lock_data, .m_lock_match = lmv_lock_match, .m_get_lustre_md = lmv_get_lustre_md, .m_free_lustre_md = lmv_free_lustre_md, .m_merge_attr = lmv_merge_attr, .m_set_open_replay_data = lmv_set_open_replay_data, .m_clear_open_replay_data = lmv_clear_open_replay_data, .m_get_remote_perm = lmv_get_remote_perm, .m_intent_getattr_async = lmv_intent_getattr_async, .m_revalidate_lock = lmv_revalidate_lock, .m_get_fid_from_lsm = lmv_get_fid_from_lsm, .m_unpackmd = lmv_unpackmd, }; static int __init lmv_init(void) { return class_register_type(&lmv_obd_ops, &lmv_md_ops, true, NULL, LUSTRE_LMV_NAME, NULL); } static void __exit lmv_exit(void) { class_unregister_type(LUSTRE_LMV_NAME); } MODULE_AUTHOR("OpenSFS, Inc. "); MODULE_DESCRIPTION("Lustre Logical Metadata Volume"); MODULE_VERSION(LUSTRE_VERSION_STRING); MODULE_LICENSE("GPL"); module_init(lmv_init); module_exit(lmv_exit);