4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 only,
8 * as published by the Free Software Foundation.
10 * This program is distributed in the hope that it will be useful, but
11 * WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * General Public License version 2 for more details (a copy is included
14 * in the LICENSE file that accompanied this code).
16 * You should have received a copy of the GNU General Public License
17 * version 2 along with this program; If not, see
18 * http://www.gnu.org/licenses/gpl-2.0.html
23 * Copyright (C) 2015, Trustees of Indiana University
25 * Copyright (c) 2017, Intel Corporation.
27 * Author: Joshua Walgenbach <jjw@iu.edu>
28 * Author: Kit Westneat <cwestnea@iu.edu>
30 * Implements the storage functionality for the nodemap configuration. Functions
31 * in this file prepare, store, and load nodemap configuration data. Targets
32 * using nodemap services should register a configuration file object. Nodemap
33 * configuration changes that need to persist should call the appropriate
34 * storage function for the data being modified.
36 * There are several index types as defined in enum nodemap_idx_type:
37 * NODEMAP_CLUSTER_IDX stores the data found on the lu_nodemap struct,
38 * like root squash and config flags, as well as
40 * NODEMAP_RANGE_IDX stores NID range information for a nodemap
41 * NODEMAP_UIDMAP_IDX stores a fs/client UID mapping pair
42 * NODEMAP_GIDMAP_IDX stores a fs/client GID mapping pair
43 * NODEMAP_GLOBAL_IDX stores whether or not nodemaps are active
46 #include <libcfs/libcfs.h>
47 #include <linux/err.h>
48 #include <linux/kernel.h>
49 #include <linux/list.h>
50 #include <linux/mutex.h>
51 #include <linux/string.h>
52 #include <linux/types.h>
53 #include <uapi/linux/lnet/lnet-types.h>
54 #include <uapi/linux/lustre/lustre_idl.h>
55 #include <uapi/linux/lustre/lustre_ioctl.h>
56 #include <uapi/linux/lustre/lustre_disk.h>
57 #include <dt_object.h>
58 #include <lu_object.h>
59 #include <lustre_net.h>
60 #include <lustre_nodemap.h>
61 #include <obd_class.h>
62 #include <obd_support.h>
63 #include "nodemap_internal.h"
65 /* list of registered nodemap index files, except MGS */
66 static LIST_HEAD(ncf_list_head);
67 static DEFINE_MUTEX(ncf_list_lock);
69 /* MGS index is different than others, others are listeners to MGS idx */
70 static struct nm_config_file *nodemap_mgs_ncf;
72 bool nodemap_mgs(void)
74 return (nodemap_mgs_ncf != NULL);
77 static void nodemap_cluster_key_init(struct nodemap_key *nk, unsigned int nm_id,
78 enum nodemap_cluster_rec_subid subid)
80 nk->nk_nodemap_id = cpu_to_le32(nm_idx_set_type(nm_id,
81 NODEMAP_CLUSTER_IDX));
82 nk->nk_cluster_subid = subid;
85 static void nodemap_cluster_rec_init(union nodemap_rec *nr,
86 const struct lu_nodemap *nodemap)
88 BUILD_BUG_ON(sizeof(nr->ncr.ncr_name) != sizeof(nodemap->nm_name));
90 strncpy(nr->ncr.ncr_name, nodemap->nm_name, sizeof(nr->ncr.ncr_name));
91 nr->ncr.ncr_squash_uid = cpu_to_le32(nodemap->nm_squash_uid);
92 nr->ncr.ncr_squash_gid = cpu_to_le32(nodemap->nm_squash_gid);
93 nr->ncr.ncr_squash_projid = cpu_to_le32(nodemap->nm_squash_projid);
95 (nodemap->nmf_trust_client_ids ?
96 NM_FL_TRUST_CLIENT_IDS : 0) |
97 (nodemap->nmf_allow_root_access ?
98 NM_FL_ALLOW_ROOT_ACCESS : 0) |
99 (nodemap->nmf_deny_unknown ?
100 NM_FL_DENY_UNKNOWN : 0) |
101 (nodemap->nmf_map_mode & NODEMAP_MAP_UID ?
103 (nodemap->nmf_map_mode & NODEMAP_MAP_GID ?
105 (nodemap->nmf_map_mode & NODEMAP_MAP_PROJID ?
106 NM_FL_MAP_PROJID : 0) |
107 (nodemap->nmf_enable_audit ?
108 NM_FL_ENABLE_AUDIT : 0) |
109 (nodemap->nmf_forbid_encryption ?
110 NM_FL_FORBID_ENCRYPT : 0);
112 (nodemap->nmf_readonly_mount ?
113 NM_FL2_READONLY_MOUNT : 0);
116 static void nodemap_cluster_roles_rec_init(union nodemap_rec *nr,
117 const struct lu_nodemap *nodemap)
119 struct nodemap_cluster_roles_rec *ncrr = &nr->ncrr;
121 memset(ncrr, 0, sizeof(struct nodemap_cluster_roles_rec));
122 ncrr->ncrr_roles = cpu_to_le64(nodemap->nmf_rbac);
125 static void nodemap_idmap_key_init(struct nodemap_key *nk, unsigned int nm_id,
126 enum nodemap_id_type id_type,
129 enum nodemap_idx_type idx_type;
131 if (id_type == NODEMAP_UID)
132 idx_type = NODEMAP_UIDMAP_IDX;
133 else if (id_type == NODEMAP_GID)
134 idx_type = NODEMAP_GIDMAP_IDX;
135 else if (id_type == NODEMAP_PROJID)
136 idx_type = NODEMAP_PROJIDMAP_IDX;
138 idx_type = NODEMAP_EMPTY_IDX;
140 nk->nk_nodemap_id = cpu_to_le32(nm_idx_set_type(nm_id, idx_type));
141 nk->nk_id_client = cpu_to_le32(id_client);
144 static void nodemap_idmap_rec_init(union nodemap_rec *nr, u32 id_fs)
146 nr->nir.nir_id_fs = cpu_to_le32(id_fs);
149 static void nodemap_range_key_init(struct nodemap_key *nk,
150 enum nodemap_idx_type type,
151 unsigned int nm_id, unsigned int rn_id)
153 nk->nk_nodemap_id = cpu_to_le32(nm_idx_set_type(nm_id, type));
154 nk->nk_range_id = cpu_to_le32(rn_id);
157 static int nodemap_range_rec_init(union nodemap_rec *nr,
158 const struct lu_nid_range *range)
160 if (range->rn_netmask) {
161 nr->nrr2.nrr_nid_prefix = range->rn_start;
162 nr->nrr2.nrr_netmask = range->rn_netmask;
164 if (NID_BYTES(&nr->nrr2.nrr_nid_prefix) >
165 sizeof(struct lnet_nid))
170 if (!nid_is_nid4(&range->rn_start) ||
171 !nid_is_nid4(&range->rn_end))
174 nid4[0] = lnet_nid_to_nid4(&range->rn_start);
175 nid4[1] = lnet_nid_to_nid4(&range->rn_end);
176 nr->nrr.nrr_start_nid = cpu_to_le64(nid4[0]);
177 nr->nrr.nrr_end_nid = cpu_to_le64(nid4[1]);
183 static void nodemap_global_key_init(struct nodemap_key *nk)
185 nk->nk_nodemap_id = cpu_to_le32(nm_idx_set_type(0, NODEMAP_GLOBAL_IDX));
189 static void nodemap_global_rec_init(union nodemap_rec *nr, bool active)
191 nr->ngr.ngr_is_active = active;
194 /* should be called with dt_write lock */
195 static void nodemap_inc_version(const struct lu_env *env,
196 struct dt_object *nodemap_idx,
199 u64 ver = dt_version_get(env, nodemap_idx);
200 dt_version_set(env, nodemap_idx, ver + 1, th);
203 enum ncfc_find_create {
207 static struct dt_object *nodemap_cache_find_create(const struct lu_env *env,
208 struct dt_device *dev,
209 struct local_oid_storage *los,
210 enum ncfc_find_create create_new)
213 struct dt_object *root_obj;
214 struct dt_object *nm_obj;
217 rc = dt_root_get(env, dev, &tfid);
219 GOTO(out, nm_obj = ERR_PTR(rc));
221 root_obj = dt_locate(env, dev, &tfid);
222 if (unlikely(IS_ERR(root_obj)))
223 GOTO(out, nm_obj = root_obj);
225 rc = dt_lookup_dir(env, root_obj, LUSTRE_NODEMAP_NAME, &tfid);
228 GOTO(out_root, nm_obj = ERR_PTR(-EROFS));
230 GOTO(out_root, nm_obj = ERR_PTR(rc));
231 } else if (dev->dd_rdonly && create_new == NCFC_CREATE_NEW) {
232 GOTO(out_root, nm_obj = ERR_PTR(-EROFS));
236 /* if loading index fails the first time, create new index */
237 if (create_new == NCFC_CREATE_NEW && rc != -ENOENT) {
238 CDEBUG(D_INFO, "removing old index, creating new one\n");
239 rc = local_object_unlink(env, dev, root_obj,
240 LUSTRE_NODEMAP_NAME);
242 /* XXX not sure the best way to get obd name. */
243 CERROR("cannot destroy nodemap index: rc = %d\n",
245 GOTO(out_root, nm_obj = ERR_PTR(rc));
250 nm_obj = local_index_find_or_create(env, los, root_obj,
252 S_IFREG | S_IRUGO | S_IWUSR,
253 &dt_nodemap_features);
254 if (IS_ERR(nm_obj)) {
255 if (PTR_ERR(nm_obj) == -EEXIST && rc != -ENOENT &&
256 los->los_last_oid < (tfid.f_oid - 1)) {
257 if (dt2lu_dev(dev)->ld_obd)
258 dt2lu_dev(dev)->ld_obd->obd_need_scrub = 1;
260 mutex_lock(&los->los_id_lock);
261 los->los_last_oid = tfid.f_oid - 1;
262 mutex_unlock(&los->los_id_lock);
267 GOTO(out_root, nm_obj);
270 if (nm_obj->do_index_ops == NULL) {
271 rc = nm_obj->do_ops->do_index_try(env, nm_obj,
272 &dt_nodemap_features);
273 /* even if loading from tgt fails, connecting to MGS will
277 dt_object_put(env, nm_obj);
279 if (create_new == NCFC_CREATE_NEW)
280 GOTO(out_root, nm_obj = ERR_PTR(rc));
282 CERROR("cannot load nodemap index from disk, creating "
283 "new index: rc = %d\n", rc);
284 create_new = NCFC_CREATE_NEW;
290 dt_object_put(env, root_obj);
295 static int nodemap_idx_insert(const struct lu_env *env,
296 struct dt_object *idx,
297 const struct nodemap_key *nk,
298 const union nodemap_rec *nr)
301 struct dt_device *dev = lu2dt_dev(idx->do_lu.lo_dev);
304 BUILD_BUG_ON(sizeof(union nodemap_rec) != 32);
306 th = dt_trans_create(env, dev);
309 GOTO(out, rc = PTR_ERR(th));
311 rc = dt_declare_insert(env, idx,
312 (const struct dt_rec *)nr,
313 (const struct dt_key *)nk, th);
317 rc = dt_declare_version_set(env, idx, th);
321 rc = dt_trans_start_local(env, dev, th);
325 dt_write_lock(env, idx, 0);
327 rc = dt_insert(env, idx, (const struct dt_rec *)nr,
328 (const struct dt_key *)nk, th);
330 nodemap_inc_version(env, idx, th);
331 dt_write_unlock(env, idx);
333 dt_trans_stop(env, dev, th);
338 static int nodemap_idx_update(const struct lu_env *env,
339 struct dt_object *idx,
340 const struct nodemap_key *nk,
341 const union nodemap_rec *nr)
344 struct dt_device *dev = lu2dt_dev(idx->do_lu.lo_dev);
347 th = dt_trans_create(env, dev);
350 GOTO(out, rc = PTR_ERR(th));
352 rc = dt_declare_delete(env, idx, (const struct dt_key *)nk, th);
356 rc = dt_declare_insert(env, idx, (const struct dt_rec *)nr,
357 (const struct dt_key *)nk, th);
361 rc = dt_declare_version_set(env, idx, th);
365 rc = dt_trans_start_local(env, dev, th);
369 dt_write_lock(env, idx, 0);
371 rc = dt_delete(env, idx, (const struct dt_key *)nk, th);
375 rc = dt_insert(env, idx, (const struct dt_rec *)nr,
376 (const struct dt_key *)nk, th);
380 nodemap_inc_version(env, idx, th);
382 dt_write_unlock(env, idx);
384 dt_trans_stop(env, dev, th);
389 static int nodemap_idx_delete(const struct lu_env *env,
390 struct dt_object *idx,
391 const struct nodemap_key *nk,
392 const union nodemap_rec *unused)
395 struct dt_device *dev = lu2dt_dev(idx->do_lu.lo_dev);
398 th = dt_trans_create(env, dev);
401 GOTO(out, rc = PTR_ERR(th));
403 rc = dt_declare_delete(env, idx, (const struct dt_key *)nk, th);
407 rc = dt_declare_version_set(env, idx, th);
411 rc = dt_trans_start_local(env, dev, th);
415 dt_write_lock(env, idx, 0);
417 rc = dt_delete(env, idx, (const struct dt_key *)nk, th);
419 nodemap_inc_version(env, idx, th);
421 dt_write_unlock(env, idx);
423 dt_trans_stop(env, dev, th);
433 static int nodemap_idx_cluster_add_update(const struct lu_nodemap *nodemap,
434 struct dt_object *idx,
435 enum nm_add_update update,
436 enum nodemap_cluster_rec_subid subid)
438 struct nodemap_key nk;
439 union nodemap_rec nr;
446 if (!nodemap_mgs()) {
447 CERROR("cannot add nodemap config to non-existing MGS.\n");
450 idx = nodemap_mgs_ncf->ncf_obj;
453 rc = lu_env_init(&env, LCT_LOCAL);
457 nodemap_cluster_key_init(&nk, nodemap->nm_id, subid);
459 case NODEMAP_CLUSTER_REC:
460 nodemap_cluster_rec_init(&nr, nodemap);
462 case NODEMAP_CLUSTER_ROLES:
463 nodemap_cluster_roles_rec_init(&nr, nodemap);
466 CWARN("%s: unknown subtype %u\n", nodemap->nm_name, subid);
467 GOTO(fini, rc = -EINVAL);
470 if (update == NM_UPDATE)
471 rc = nodemap_idx_update(&env, idx, &nk, &nr);
473 rc = nodemap_idx_insert(&env, idx, &nk, &nr);
480 int nodemap_idx_nodemap_add(const struct lu_nodemap *nodemap)
482 return nodemap_idx_cluster_add_update(nodemap, NULL,
483 NM_ADD, NODEMAP_CLUSTER_REC);
486 int nodemap_idx_nodemap_update(const struct lu_nodemap *nodemap)
488 return nodemap_idx_cluster_add_update(nodemap, NULL,
489 NM_UPDATE, NODEMAP_CLUSTER_REC);
492 int nodemap_idx_nodemap_del(const struct lu_nodemap *nodemap)
495 struct lu_idmap *idmap;
496 struct lu_idmap *temp;
497 struct lu_nid_range *range;
498 struct lu_nid_range *range_temp;
499 struct nodemap_key nk;
505 if (!nodemap_mgs()) {
506 CERROR("cannot add nodemap config to non-existing MGS.\n");
510 rc = lu_env_init(&env, LCT_LOCAL);
514 nodemap_cluster_key_init(&nk, nodemap->nm_id, NODEMAP_CLUSTER_ROLES);
515 rc2 = nodemap_idx_delete(&env, nodemap_mgs_ncf->ncf_obj, &nk, NULL);
516 if (rc2 < 0 && rc2 != -ENOENT)
519 root = nodemap->nm_fs_to_client_uidmap;
520 rbtree_postorder_for_each_entry_safe(idmap, temp, &root,
522 nodemap_idmap_key_init(&nk, nodemap->nm_id, NODEMAP_UID,
524 rc2 = nodemap_idx_delete(&env, nodemap_mgs_ncf->ncf_obj,
530 root = nodemap->nm_client_to_fs_gidmap;
531 rbtree_postorder_for_each_entry_safe(idmap, temp, &root,
533 nodemap_idmap_key_init(&nk, nodemap->nm_id, NODEMAP_GID,
535 rc2 = nodemap_idx_delete(&env, nodemap_mgs_ncf->ncf_obj,
541 root = nodemap->nm_client_to_fs_projidmap;
542 rbtree_postorder_for_each_entry_safe(idmap, temp, &root,
544 nodemap_idmap_key_init(&nk, nodemap->nm_id, NODEMAP_PROJID,
546 rc2 = nodemap_idx_delete(&env, nodemap_mgs_ncf->ncf_obj,
552 list_for_each_entry_safe(range, range_temp, &nodemap->nm_ranges,
554 enum nodemap_idx_type type;
556 type = range->rn_netmask ? NODEMAP_NID_MASK_IDX :
558 nodemap_range_key_init(&nk, type, nodemap->nm_id, range->rn_id);
559 rc2 = nodemap_idx_delete(&env, nodemap_mgs_ncf->ncf_obj,
565 nodemap_cluster_key_init(&nk, nodemap->nm_id, NODEMAP_CLUSTER_REC);
566 rc2 = nodemap_idx_delete(&env, nodemap_mgs_ncf->ncf_obj, &nk, NULL);
575 int nodemap_idx_cluster_roles_add(const struct lu_nodemap *nodemap)
577 return nodemap_idx_cluster_add_update(nodemap, NULL, NM_ADD,
578 NODEMAP_CLUSTER_ROLES);
581 int nodemap_idx_cluster_roles_update(const struct lu_nodemap *nodemap)
583 return nodemap_idx_cluster_add_update(nodemap, NULL, NM_UPDATE,
584 NODEMAP_CLUSTER_ROLES);
587 int nodemap_idx_cluster_roles_del(const struct lu_nodemap *nodemap)
589 struct nodemap_key nk;
595 if (!nodemap_mgs()) {
596 CERROR("cannot add nodemap config to non-existing MGS.\n");
600 rc = lu_env_init(&env, LCT_LOCAL);
604 nodemap_cluster_key_init(&nk, nodemap->nm_id, NODEMAP_CLUSTER_ROLES);
605 rc = nodemap_idx_delete(&env, nodemap_mgs_ncf->ncf_obj, &nk, NULL);
611 int nodemap_idx_range_add(const struct lu_nid_range *range)
613 struct nodemap_key nk;
614 union nodemap_rec nr;
619 if (!nodemap_mgs()) {
620 CERROR("cannot add nodemap config to non-existing MGS.\n");
624 rc = lu_env_init(&env, LCT_LOCAL);
628 nodemap_range_key_init(&nk, range->rn_netmask ? NODEMAP_NID_MASK_IDX :
630 range->rn_nodemap->nm_id, range->rn_id);
631 rc = nodemap_range_rec_init(&nr, range);
635 rc = nodemap_idx_insert(&env, nodemap_mgs_ncf->ncf_obj, &nk, &nr);
642 int nodemap_idx_range_del(const struct lu_nid_range *range)
644 struct nodemap_key nk;
649 if (!nodemap_mgs()) {
650 CERROR("cannot del nodemap config from non-existing MGS.\n");
654 rc = lu_env_init(&env, LCT_LOCAL);
658 nodemap_range_key_init(&nk, range->rn_netmask ? NODEMAP_NID_MASK_IDX :
660 range->rn_nodemap->nm_id, range->rn_id);
661 rc = nodemap_idx_delete(&env, nodemap_mgs_ncf->ncf_obj, &nk, NULL);
667 int nodemap_idx_idmap_add(const struct lu_nodemap *nodemap,
668 enum nodemap_id_type id_type,
671 struct nodemap_key nk;
672 union nodemap_rec nr;
677 if (!nodemap_mgs()) {
678 CERROR("cannot add nodemap config to non-existing MGS.\n");
682 rc = lu_env_init(&env, LCT_LOCAL);
686 nodemap_idmap_key_init(&nk, nodemap->nm_id, id_type, map[0]);
687 nodemap_idmap_rec_init(&nr, map[1]);
689 rc = nodemap_idx_insert(&env, nodemap_mgs_ncf->ncf_obj, &nk, &nr);
695 int nodemap_idx_idmap_del(const struct lu_nodemap *nodemap,
696 enum nodemap_id_type id_type,
699 struct nodemap_key nk;
704 if (!nodemap_mgs()) {
705 CERROR("cannot add nodemap config to non-existing MGS.\n");
709 rc = lu_env_init(&env, LCT_LOCAL);
713 nodemap_idmap_key_init(&nk, nodemap->nm_id, id_type, map[0]);
715 rc = nodemap_idx_delete(&env, nodemap_mgs_ncf->ncf_obj, &nk, NULL);
721 static int nodemap_idx_global_add_update(bool value, enum nm_add_update update)
723 struct nodemap_key nk;
724 union nodemap_rec nr;
729 if (!nodemap_mgs()) {
730 CERROR("cannot add nodemap config to non-existing MGS.\n");
734 rc = lu_env_init(&env, LCT_LOCAL);
738 nodemap_global_key_init(&nk);
739 nodemap_global_rec_init(&nr, value);
741 if (update == NM_UPDATE)
742 rc = nodemap_idx_update(&env, nodemap_mgs_ncf->ncf_obj,
745 rc = nodemap_idx_insert(&env, nodemap_mgs_ncf->ncf_obj,
753 int nodemap_idx_nodemap_activate(bool value)
755 return nodemap_idx_global_add_update(value, NM_UPDATE);
758 static enum nodemap_idx_type nodemap_get_key_type(const struct nodemap_key *key)
762 nodemap_id = le32_to_cpu(key->nk_nodemap_id);
763 return nm_idx_get_type(nodemap_id);
766 static int nodemap_get_key_subtype(const struct nodemap_key *key)
768 enum nodemap_idx_type type = nodemap_get_key_type(key);
770 return type == NODEMAP_CLUSTER_IDX ? key->nk_cluster_subid : -1;
773 static int nodemap_cluster_rec_helper(struct nodemap_config *config,
775 const union nodemap_rec *rec,
776 struct lu_nodemap **recent_nodemap)
778 struct lu_nodemap *nodemap, *old_nm;
779 enum nm_flag_bits flags;
780 enum nm_flag2_bits flags2;
782 nodemap = cfs_hash_lookup(config->nmc_nodemap_hash, rec->ncr.ncr_name);
783 if (nodemap == NULL) {
784 if (nodemap_id == LUSTRE_NODEMAP_DEFAULT_ID)
785 nodemap = nodemap_create(rec->ncr.ncr_name, config, 1);
787 nodemap = nodemap_create(rec->ncr.ncr_name, config, 0);
789 return PTR_ERR(nodemap);
791 /* we need to override the local ID with the saved ID */
792 nodemap->nm_id = nodemap_id;
793 if (nodemap_id > config->nmc_nodemap_highest_id)
794 config->nmc_nodemap_highest_id = nodemap_id;
796 } else if (nodemap->nm_id != nodemap_id) {
797 nodemap_putref(nodemap);
801 nodemap->nm_squash_uid = le32_to_cpu(rec->ncr.ncr_squash_uid);
802 nodemap->nm_squash_gid = le32_to_cpu(rec->ncr.ncr_squash_gid);
803 nodemap->nm_squash_projid = le32_to_cpu(rec->ncr.ncr_squash_projid);
805 flags = rec->ncr.ncr_flags;
806 nodemap->nmf_allow_root_access = flags & NM_FL_ALLOW_ROOT_ACCESS;
807 nodemap->nmf_trust_client_ids = flags & NM_FL_TRUST_CLIENT_IDS;
808 nodemap->nmf_deny_unknown = flags & NM_FL_DENY_UNKNOWN;
809 nodemap->nmf_map_mode =
810 (flags & NM_FL_MAP_UID ? NODEMAP_MAP_UID : 0) |
811 (flags & NM_FL_MAP_GID ? NODEMAP_MAP_GID : 0) |
812 (flags & NM_FL_MAP_PROJID ? NODEMAP_MAP_PROJID : 0);
813 if (nodemap->nmf_map_mode == NODEMAP_MAP_BOTH_LEGACY)
814 nodemap->nmf_map_mode = NODEMAP_MAP_BOTH;
815 nodemap->nmf_enable_audit = flags & NM_FL_ENABLE_AUDIT;
816 nodemap->nmf_forbid_encryption = flags & NM_FL_FORBID_ENCRYPT;
817 flags2 = rec->ncr.ncr_flags2;
818 nodemap->nmf_readonly_mount = flags2 & NM_FL2_READONLY_MOUNT;
819 /* by default, and in the absence of cluster_roles, grant all roles */
820 nodemap->nmf_rbac = NODEMAP_RBAC_ALL;
822 /* The fileset should be saved otherwise it will be empty
823 * every time in case of "NODEMAP_CLUSTER_IDX".
825 mutex_lock(&active_config_lock);
826 old_nm = nodemap_lookup(rec->ncr.ncr_name);
827 if (!IS_ERR(old_nm) && old_nm->nm_fileset[0] != '\0')
828 strscpy(nodemap->nm_fileset, old_nm->nm_fileset,
829 sizeof(nodemap->nm_fileset));
830 mutex_unlock(&active_config_lock);
832 nodemap_putref(old_nm);
834 if (*recent_nodemap == NULL) {
835 *recent_nodemap = nodemap;
836 INIT_LIST_HEAD(&nodemap->nm_list);
838 list_add(&nodemap->nm_list, &(*recent_nodemap)->nm_list);
840 nodemap_putref(nodemap);
845 static int nodemap_cluster_roles_helper(struct lu_nodemap *nodemap,
846 const union nodemap_rec *rec)
848 nodemap->nmf_rbac = le64_to_cpu(rec->ncrr.ncrr_roles);
854 * Process a key/rec pair and modify the new configuration.
856 * \param config configuration to update with this key/rec data
857 * \param key key of the record that was loaded
858 * \param rec record that was loaded
859 * \param recent_nodemap last referenced nodemap
860 * \retval type of record processed, see enum #nodemap_idx_type
861 * \retval -ENOENT range or map loaded before nodemap record
862 * \retval -EINVAL duplicate nodemap cluster records found with
863 * different IDs, or nodemap has invalid name
866 static int nodemap_process_keyrec(struct nodemap_config *config,
867 const struct nodemap_key *key,
868 const union nodemap_rec *rec,
869 struct lu_nodemap **recent_nodemap)
871 struct lu_nodemap *nodemap = NULL;
872 enum nodemap_idx_type type;
873 enum nodemap_id_type id_type;
874 struct lnet_nid nid[2];
882 BUILD_BUG_ON(sizeof(union nodemap_rec) != 32);
884 nodemap_id = le32_to_cpu(key->nk_nodemap_id);
885 type = nodemap_get_key_type(key);
886 subtype = nodemap_get_key_subtype(key);
887 nodemap_id = nm_idx_set_type(nodemap_id, 0);
889 CDEBUG(D_INFO, "found config entry, nm_id %d type %d subtype %d\n",
890 nodemap_id, type, subtype);
892 /* find the correct nodemap in the load list */
893 if (type == NODEMAP_RANGE_IDX || type == NODEMAP_NID_MASK_IDX ||
894 type == NODEMAP_UIDMAP_IDX || type == NODEMAP_GIDMAP_IDX ||
895 type == NODEMAP_PROJIDMAP_IDX ||
896 (type == NODEMAP_CLUSTER_IDX && subtype != NODEMAP_CLUSTER_REC)) {
897 struct lu_nodemap *tmp = NULL;
899 nodemap = *recent_nodemap;
902 GOTO(out, rc = -ENOENT);
904 if (nodemap->nm_id != nodemap_id) {
905 list_for_each_entry(tmp, &nodemap->nm_list, nm_list)
906 if (tmp->nm_id == nodemap_id) {
911 if (nodemap->nm_id != nodemap_id)
912 GOTO(out, rc = -ENOENT);
915 /* update most recently used nodemap if necessay */
916 if (nodemap != *recent_nodemap)
917 *recent_nodemap = nodemap;
921 case NODEMAP_EMPTY_IDX:
923 CWARN("Found nodemap config record without type field, "
924 " nodemap_id=%d. nodemap config file corrupt?\n",
927 case NODEMAP_CLUSTER_IDX:
928 switch (nodemap_get_key_subtype(key)) {
929 case NODEMAP_CLUSTER_REC:
930 rc = nodemap_cluster_rec_helper(config, nodemap_id, rec,
935 case NODEMAP_CLUSTER_ROLES:
936 rc = nodemap_cluster_roles_helper(nodemap, rec);
941 CWARN("%s: ignoring keyrec of type %d with subtype %u\n",
942 nodemap->nm_name, NODEMAP_CLUSTER_IDX,
943 nodemap_get_key_subtype(key));
947 case NODEMAP_RANGE_IDX:
948 lnet_nid4_to_nid(le64_to_cpu(rec->nrr.nrr_start_nid), &nid[0]);
949 lnet_nid4_to_nid(le64_to_cpu(rec->nrr.nrr_end_nid), &nid[1]);
950 rc = nodemap_add_range_helper(config, nodemap, nid, 0,
951 le32_to_cpu(key->nk_range_id));
955 case NODEMAP_NID_MASK_IDX:
956 nid[0] = rec->nrr2.nrr_nid_prefix;
957 nid[1] = rec->nrr2.nrr_nid_prefix;
958 rc = nodemap_add_range_helper(config, nodemap, nid,
959 rec->nrr2.nrr_netmask,
960 le32_to_cpu(key->nk_range_id));
964 case NODEMAP_UIDMAP_IDX:
965 case NODEMAP_GIDMAP_IDX:
966 case NODEMAP_PROJIDMAP_IDX:
967 map[0] = le32_to_cpu(key->nk_id_client);
968 map[1] = le32_to_cpu(rec->nir.nir_id_fs);
970 if (type == NODEMAP_UIDMAP_IDX)
971 id_type = NODEMAP_UID;
972 else if (type == NODEMAP_GIDMAP_IDX)
973 id_type = NODEMAP_GID;
974 else if (type == NODEMAP_PROJIDMAP_IDX)
975 id_type = NODEMAP_PROJID;
977 GOTO(out, rc = -EINVAL);
979 rc = nodemap_add_idmap_helper(nodemap, id_type, map);
983 case NODEMAP_GLOBAL_IDX:
984 switch (key->nk_unused) {
986 config->nmc_nodemap_is_active = rec->ngr.ngr_is_active;
989 CWARN("%s: ignoring keyrec of type %d with subtype %u\n",
991 (*recent_nodemap)->nm_name : "nodemap",
992 NODEMAP_GLOBAL_IDX, key->nk_unused);
997 CWARN("%s: ignoring key %u:%u for unknown type %u\n",
998 recent_nodemap ? (*recent_nodemap)->nm_name : "nodemap",
999 key->nk_nodemap_id & 0x0FFFFFFF, key->nk_unused, type);
1011 enum nm_config_passes {
1012 NM_READ_CLUSTERS = 0,
1013 NM_READ_ATTRIBUTES = 1,
1016 static int nodemap_load_entries(const struct lu_env *env,
1017 struct dt_object *nodemap_idx)
1019 const struct dt_it_ops *iops;
1021 struct lu_nodemap *recent_nodemap = NULL;
1022 struct nodemap_config *new_config = NULL;
1024 bool activate_nodemap = false;
1025 bool loaded_global_idx = false;
1026 enum nm_config_passes cur_pass = NM_READ_CLUSTERS;
1031 iops = &nodemap_idx->do_index_ops->dio_it;
1033 dt_read_lock(env, nodemap_idx, 0);
1034 it = iops->init(env, nodemap_idx, 0);
1036 GOTO(out, rc = PTR_ERR(it));
1038 rc = iops->load(env, it, hash);
1040 GOTO(out_iops_fini, rc);
1042 /* rc == 0 means we need to advance to record */
1044 rc = iops->next(env, it);
1047 GOTO(out_iops_put, rc);
1048 /* rc > 0 is eof, will be checked in while below */
1050 /* rc == 1, we found initial record and can process below */
1054 new_config = nodemap_config_alloc();
1055 if (IS_ERR(new_config)) {
1056 rc = PTR_ERR(new_config);
1058 GOTO(out_iops_put, rc);
1061 /* rc > 0 is eof, check initial iops->next here as well */
1063 struct nodemap_key *key;
1064 union nodemap_rec rec;
1065 enum nodemap_idx_type key_type;
1068 key = (struct nodemap_key *)iops->key(env, it);
1069 key_type = nodemap_get_key_type((struct nodemap_key *)key);
1070 sub_type = nodemap_get_key_subtype((struct nodemap_key *)key);
1071 if ((cur_pass == NM_READ_CLUSTERS &&
1072 key_type == NODEMAP_CLUSTER_IDX &&
1073 sub_type == NODEMAP_CLUSTER_REC) ||
1074 (cur_pass == NM_READ_ATTRIBUTES &&
1075 (key_type != NODEMAP_CLUSTER_IDX ||
1076 sub_type != NODEMAP_CLUSTER_REC) &&
1077 key_type != NODEMAP_EMPTY_IDX)) {
1078 rc = iops->rec(env, it, (struct dt_rec *)&rec, 0);
1079 if (rc != -ESTALE) {
1081 GOTO(out_nodemap_config, rc);
1082 rc = nodemap_process_keyrec(new_config, key, &rec,
1085 GOTO(out_nodemap_config, rc);
1086 if (rc == NODEMAP_GLOBAL_IDX)
1087 loaded_global_idx = true;
1092 rc = iops->next(env, it);
1093 while (rc == -ESTALE);
1095 /* move to second pass */
1096 if (rc > 0 && cur_pass == NM_READ_CLUSTERS) {
1097 cur_pass = NM_READ_ATTRIBUTES;
1098 rc = iops->load(env, it, 0);
1100 rc = iops->next(env, it);
1113 nodemap_config_dealloc(new_config);
1115 /* creating new default needs to be done outside dt read lock */
1116 activate_nodemap = true;
1120 iops->fini(env, it);
1122 dt_read_unlock(env, nodemap_idx);
1125 CWARN("%s: failed to load nodemap configuration: rc = %d\n",
1126 nodemap_idx->do_lu.lo_dev->ld_obd->obd_name, rc);
1128 if (!activate_nodemap)
1131 if (new_config->nmc_default_nodemap == NULL) {
1132 /* new MGS won't have a default nm on disk, so create it here */
1133 struct lu_nodemap *nodemap =
1134 nodemap_create(DEFAULT_NODEMAP, new_config, 1);
1135 if (IS_ERR(nodemap)) {
1136 rc = PTR_ERR(nodemap);
1138 rc = nodemap_idx_cluster_add_update(
1139 new_config->nmc_default_nodemap,
1141 NM_ADD, NODEMAP_CLUSTER_REC);
1142 nodemap_putref(new_config->nmc_default_nodemap);
1146 /* new nodemap config won't have an active/inactive record */
1147 if (rc == 0 && loaded_global_idx == false) {
1148 struct nodemap_key nk;
1149 union nodemap_rec nr;
1151 nodemap_global_key_init(&nk);
1152 nodemap_global_rec_init(&nr, false);
1153 rc = nodemap_idx_insert(env, nodemap_idx, &nk, &nr);
1157 nodemap_config_set_active(new_config);
1159 nodemap_config_dealloc(new_config);
1165 * Step through active config and write to disk.
1167 static struct dt_object *
1168 nodemap_save_config_cache(const struct lu_env *env,
1169 struct dt_device *dev,
1170 struct local_oid_storage *los)
1172 struct dt_object *o;
1173 struct lu_nodemap *nodemap;
1174 struct lu_nodemap *nm_tmp;
1175 struct lu_nid_range *range;
1176 struct lu_nid_range *range_temp;
1177 struct lu_idmap *idmap;
1178 struct lu_idmap *id_tmp;
1179 struct rb_root root;
1180 struct nodemap_key nk;
1181 union nodemap_rec nr;
1182 LIST_HEAD(nodemap_list_head);
1187 /* create a new index file to fill with active config */
1188 o = nodemap_cache_find_create(env, dev, los, NCFC_CREATE_NEW);
1192 mutex_lock(&active_config_lock);
1194 /* convert hash to list so we don't spin */
1195 cfs_hash_for_each_safe(active_config->nmc_nodemap_hash,
1196 nm_hash_list_cb, &nodemap_list_head);
1198 list_for_each_entry_safe(nodemap, nm_tmp, &nodemap_list_head, nm_list) {
1199 nodemap_cluster_key_init(&nk, nodemap->nm_id,
1200 NODEMAP_CLUSTER_REC);
1201 nodemap_cluster_rec_init(&nr, nodemap);
1203 rc2 = nodemap_idx_insert(env, o, &nk, &nr);
1209 /* only insert NODEMAP_CLUSTER_ROLES idx in saved config cache
1210 * if nmf_rbac is not default value NODEMAP_RBAC_ALL
1212 if (nodemap->nmf_rbac != NODEMAP_RBAC_ALL) {
1213 nodemap_cluster_key_init(&nk, nodemap->nm_id,
1214 NODEMAP_CLUSTER_ROLES);
1215 nodemap_cluster_roles_rec_init(&nr, nodemap);
1216 rc2 = nodemap_idx_insert(env, o, &nk, &nr);
1221 down_read(&active_config->nmc_range_tree_lock);
1222 list_for_each_entry_safe(range, range_temp, &nodemap->nm_ranges,
1224 enum nodemap_idx_type type;
1226 type = range->rn_netmask ? NODEMAP_NID_MASK_IDX :
1228 nodemap_range_key_init(&nk, type, nodemap->nm_id,
1230 rc2 = nodemap_range_rec_init(&nr, range);
1235 rc2 = nodemap_idx_insert(env, o, &nk, &nr);
1239 up_read(&active_config->nmc_range_tree_lock);
1241 /* we don't need to take nm_idmap_lock because active config
1242 * lock prevents changes from happening to nodemaps
1244 root = nodemap->nm_client_to_fs_uidmap;
1245 rbtree_postorder_for_each_entry_safe(idmap, id_tmp, &root,
1247 nodemap_idmap_key_init(&nk, nodemap->nm_id, NODEMAP_UID,
1249 nodemap_idmap_rec_init(&nr, idmap->id_fs);
1250 rc2 = nodemap_idx_insert(env, o, &nk, &nr);
1255 root = nodemap->nm_client_to_fs_gidmap;
1256 rbtree_postorder_for_each_entry_safe(idmap, id_tmp, &root,
1258 nodemap_idmap_key_init(&nk, nodemap->nm_id, NODEMAP_GID,
1260 nodemap_idmap_rec_init(&nr, idmap->id_fs);
1261 rc2 = nodemap_idx_insert(env, o, &nk, &nr);
1266 root = nodemap->nm_client_to_fs_projidmap;
1267 rbtree_postorder_for_each_entry_safe(idmap, id_tmp, &root,
1269 nodemap_idmap_key_init(&nk, nodemap->nm_id,
1272 nodemap_idmap_rec_init(&nr, idmap->id_fs);
1273 rc2 = nodemap_idx_insert(env, o, &nk, &nr);
1278 nodemap_global_key_init(&nk);
1279 nodemap_global_rec_init(&nr, active_config->nmc_nodemap_is_active);
1280 rc2 = nodemap_idx_insert(env, o, &nk, &nr);
1284 mutex_unlock(&active_config_lock);
1287 dt_object_put(env, o);
1294 static void nodemap_save_all_caches(void)
1296 struct nm_config_file *ncf;
1300 /* recreating nodemap cache requires fld_thread_key be in env */
1301 rc = lu_env_init(&env, LCT_MD_THREAD | LCT_DT_THREAD | LCT_MG_THREAD);
1303 CWARN("cannot init env for nodemap config: rc = %d\n", rc);
1307 mutex_lock(&ncf_list_lock);
1308 list_for_each_entry(ncf, &ncf_list_head, ncf_list) {
1309 struct dt_device *dev = lu2dt_dev(ncf->ncf_obj->do_lu.lo_dev);
1310 struct obd_device *obd = ncf->ncf_obj->do_lu.lo_dev->ld_obd;
1311 struct dt_object *o;
1313 /* put current config file so save conf can rewrite it */
1314 dt_object_put_nocache(&env, ncf->ncf_obj);
1315 ncf->ncf_obj = NULL;
1317 o = nodemap_save_config_cache(&env, dev, ncf->ncf_los);
1319 CWARN("%s: error writing to nodemap config: rc = %d\n",
1324 mutex_unlock(&ncf_list_lock);
1329 /* tracks if config still needs to be loaded, either from disk or network */
1330 /* 0: not loaded yet
1331 * 1: successfully loaded
1332 * -1: loading in progress
1334 static int nodemap_config_loaded;
1335 static DEFINE_MUTEX(nodemap_config_loaded_lock);
1337 bool nodemap_loading(void)
1339 return (nodemap_config_loaded == -1);
1342 void nodemap_config_set_loading_mgc(bool loading)
1344 mutex_lock(&nodemap_config_loaded_lock);
1345 nodemap_config_loaded = loading ? -1 : 0;
1346 mutex_unlock(&nodemap_config_loaded_lock);
1348 EXPORT_SYMBOL(nodemap_config_set_loading_mgc);
1351 * Ensures that configs loaded over the wire are prioritized over those loaded
1354 * \param config config to set as the active config
1356 void nodemap_config_set_active_mgc(struct nodemap_config *config)
1358 mutex_lock(&nodemap_config_loaded_lock);
1359 nodemap_config_set_active(config);
1360 nodemap_config_loaded = 1;
1361 nodemap_save_all_caches();
1362 mutex_unlock(&nodemap_config_loaded_lock);
1364 EXPORT_SYMBOL(nodemap_config_set_active_mgc);
1367 * Register a dt_object representing the config index file. This should be
1368 * called by targets in order to load the nodemap configuration from disk. The
1369 * dt_object should be created with local_index_find_or_create and the index
1370 * features should be enabled with do_index_try.
1372 * \param obj dt_object returned by local_index_find_or_create
1374 * \retval on success: nm_config_file handle for later deregistration
1375 * \retval -ENOMEM memory allocation failure
1376 * \retval -ENOENT error loading nodemap config
1377 * \retval -EINVAL error loading nodemap config
1378 * \retval -EEXIST nodemap config already registered for MGS
1380 struct nm_config_file *nm_config_file_register_mgs(const struct lu_env *env,
1381 struct dt_object *obj,
1382 struct local_oid_storage *los)
1384 struct nm_config_file *ncf;
1389 GOTO(out, ncf = ERR_PTR(-EEXIST));
1393 GOTO(out, ncf = ERR_PTR(-ENOMEM));
1395 /* if loading from cache, prevent activation of MGS config until cache
1396 * loading is done, so disk config is overwritten by MGS config.
1398 mutex_lock(&nodemap_config_loaded_lock);
1399 nodemap_config_loaded = -1;
1400 rc = nodemap_load_entries(env, obj);
1401 nodemap_config_loaded = !rc;
1402 mutex_unlock(&nodemap_config_loaded_lock);
1406 GOTO(out, ncf = ERR_PTR(rc));
1409 lu_object_get(&obj->do_lu);
1414 nodemap_mgs_ncf = ncf;
1419 EXPORT_SYMBOL(nm_config_file_register_mgs);
1421 struct nm_config_file *nm_config_file_register_tgt(const struct lu_env *env,
1422 struct dt_device *dev,
1423 struct local_oid_storage *los)
1425 struct nm_config_file *ncf;
1426 struct dt_object *config_obj = NULL;
1431 RETURN(ERR_PTR(-ENOMEM));
1433 /* don't load from cache if config already loaded */
1434 mutex_lock(&nodemap_config_loaded_lock);
1435 if (nodemap_config_loaded < 1) {
1436 config_obj = nodemap_cache_find_create(env, dev, los, 0);
1437 if (IS_ERR(config_obj)) {
1438 rc = PTR_ERR(config_obj);
1440 nodemap_config_loaded = -1;
1441 rc = nodemap_load_entries(env, config_obj);
1443 nodemap_config_loaded = !rc;
1445 mutex_unlock(&nodemap_config_loaded_lock);
1449 /* sync on disk caches w/ loaded config in memory, ncf_obj may change */
1451 config_obj = nodemap_save_config_cache(env, dev, los);
1452 if (IS_ERR(config_obj))
1453 GOTO(out_ncf, rc = PTR_ERR(config_obj));
1456 ncf->ncf_obj = config_obj;
1459 mutex_lock(&ncf_list_lock);
1460 list_add(&ncf->ncf_list, &ncf_list_head);
1461 mutex_unlock(&ncf_list_lock);
1466 RETURN(ERR_PTR(rc));
1471 EXPORT_SYMBOL(nm_config_file_register_tgt);
1474 * Deregister a nm_config_file. Should be called by targets during cleanup.
1476 * \param ncf config file to deregister
1478 void nm_config_file_deregister_mgs(const struct lu_env *env,
1479 struct nm_config_file *ncf)
1482 LASSERT(nodemap_mgs_ncf == ncf);
1484 nodemap_mgs_ncf = NULL;
1486 dt_object_put(env, ncf->ncf_obj);
1492 EXPORT_SYMBOL(nm_config_file_deregister_mgs);
1494 void nm_config_file_deregister_tgt(const struct lu_env *env,
1495 struct nm_config_file *ncf)
1502 mutex_lock(&ncf_list_lock);
1503 list_del(&ncf->ncf_list);
1504 mutex_unlock(&ncf_list_lock);
1507 dt_object_put(env, ncf->ncf_obj);
1513 EXPORT_SYMBOL(nm_config_file_deregister_tgt);
1515 int nodemap_process_idx_pages(struct nodemap_config *config, union lu_page *lip,
1516 struct lu_nodemap **recent_nodemap)
1518 struct nodemap_key *key;
1519 union nodemap_rec *rec;
1524 int size = dt_nodemap_features.dif_keysize_max +
1525 dt_nodemap_features.dif_recsize_max;
1528 for (j = 0; j < LU_PAGE_COUNT; j++) {
1529 if (lip->lp_idx.lip_magic != LIP_MAGIC)
1532 /* get and process keys and records from page */
1533 for (k = 0; k < lip->lp_idx.lip_nr; k++) {
1534 entry = lip->lp_idx.lip_entries + k * size;
1535 key = (struct nodemap_key *)entry;
1537 entry += dt_nodemap_features.dif_keysize_max;
1538 rec = (union nodemap_rec *)entry;
1540 rc = nodemap_process_keyrec(config, key, rec,
1551 EXPORT_SYMBOL(nodemap_process_idx_pages);
1553 static int nodemap_page_build(const struct lu_env *env, struct dt_object *obj,
1554 union lu_page *lp, size_t bytes,
1555 const struct dt_it_ops *iops,
1556 struct dt_it *it, __u32 attr, void *arg)
1558 struct idx_info *ii = (struct idx_info *)arg;
1559 struct lu_idxpage *lip = &lp->lp_idx;
1561 size_t size = ii->ii_keysize + ii->ii_recsize;
1565 if (bytes < LIP_HDR_SIZE)
1568 /* initialize the header of the new container */
1569 memset(lip, 0, LIP_HDR_SIZE);
1570 lip->lip_magic = LIP_MAGIC;
1571 bytes -= LIP_HDR_SIZE;
1573 entry = lip->lip_entries;
1575 char *tmp_entry = entry;
1578 enum nodemap_idx_type key_type;
1581 /* fetch 64-bit hash value */
1582 hash = iops->store(env, it);
1583 ii->ii_hash_end = hash;
1585 if (CFS_FAIL_CHECK(OBD_FAIL_OBD_IDX_READ_BREAK)) {
1586 if (lip->lip_nr != 0)
1591 if (lip->lip_nr == 0)
1592 GOTO(out, rc = -EINVAL);
1596 key = iops->key(env, it);
1597 key_type = nodemap_get_key_type((struct nodemap_key *)key);
1598 sub_type = nodemap_get_key_subtype((struct nodemap_key *)key);
1600 /* on the first pass, get only the cluster types. On second
1601 * pass, get all the rest */
1602 if ((ii->ii_attrs == NM_READ_CLUSTERS &&
1603 key_type == NODEMAP_CLUSTER_IDX &&
1604 sub_type == NODEMAP_CLUSTER_REC) ||
1605 (ii->ii_attrs == NM_READ_ATTRIBUTES &&
1606 (key_type != NODEMAP_CLUSTER_IDX ||
1607 sub_type != NODEMAP_CLUSTER_REC) &&
1608 key_type != NODEMAP_EMPTY_IDX)) {
1609 memcpy(tmp_entry, key, ii->ii_keysize);
1610 tmp_entry += ii->ii_keysize;
1612 /* and finally the record */
1613 rc = iops->rec(env, it, (struct dt_rec *)tmp_entry,
1615 if (rc != -ESTALE) {
1619 /* hash/key/record successfully copied! */
1621 if (unlikely(lip->lip_nr == 1 &&
1623 ii->ii_hash_start = hash;
1625 entry = tmp_entry + ii->ii_recsize;
1630 /* move on to the next record */
1632 rc = iops->next(env, it);
1633 } while (rc == -ESTALE);
1635 /* move to second pass */
1636 if (rc > 0 && ii->ii_attrs == NM_READ_CLUSTERS) {
1637 ii->ii_attrs = NM_READ_ATTRIBUTES;
1638 rc = iops->load(env, it, 0);
1640 rc = iops->next(env, it);
1651 if (rc >= 0 && lip->lip_nr > 0)
1652 /* one more container */
1655 /* no more entries */
1656 ii->ii_hash_end = II_END_OFF;
1660 int nodemap_index_read(struct lu_env *env, struct nm_config_file *ncf,
1661 struct idx_info *ii, const struct lu_rdpg *rdpg)
1663 struct dt_object *nodemap_idx = ncf->ncf_obj;
1667 ii->ii_keysize = dt_nodemap_features.dif_keysize_max;
1668 ii->ii_recsize = dt_nodemap_features.dif_recsize_max;
1670 dt_read_lock(env, nodemap_idx, 0);
1671 version = dt_version_get(env, nodemap_idx);
1672 if (rdpg->rp_hash != 0 && ii->ii_version != version) {
1673 CDEBUG(D_INFO, "nodemap config changed inflight, old %llu, new %llu\n",
1676 ii->ii_hash_end = 0;
1678 rc = dt_index_walk(env, nodemap_idx, rdpg, nodemap_page_build,
1680 CDEBUG(D_INFO, "walked index, hashend %llx\n", ii->ii_hash_end);
1684 ii->ii_version = version;
1687 * For partial lu_idxpage filling of the end system page,
1688 * init the header of the remain lu_idxpages.
1691 dt_index_page_adjust(rdpg->rp_pages, rdpg->rp_npages,
1694 dt_read_unlock(env, nodemap_idx);
1697 EXPORT_SYMBOL(nodemap_index_read);
1700 * Returns the current nodemap configuration to MGC by walking the nodemap
1701 * config index and storing it in the response buffer.
1703 * \param req incoming MGS_CONFIG_READ request
1705 * \retval -EINVAL malformed request
1706 * \retval -ENOTCONN client evicted/reconnected already
1707 * \retval -ETIMEDOUT client timeout or network error
1710 int nodemap_get_config_req(struct obd_device *mgs_obd,
1711 struct ptlrpc_request *req)
1713 const struct ptlrpc_bulk_frag_ops *frag_ops = &ptlrpc_bulk_kiov_pin_ops;
1714 struct mgs_config_body *body;
1715 struct mgs_config_res *res;
1716 struct lu_rdpg rdpg;
1717 struct idx_info nodemap_ii;
1718 struct ptlrpc_bulk_desc *desc;
1719 struct tg_export_data *rqexp_ted = &req->rq_export->exp_target_data;
1725 body = req_capsule_client_get(&req->rq_pill, &RMF_MGS_CONFIG_BODY);
1729 if (body->mcb_type != MGS_CFG_T_NODEMAP)
1732 rdpg.rp_count = (body->mcb_units << body->mcb_bits);
1733 rdpg.rp_npages = (rdpg.rp_count + PAGE_SIZE - 1) >>
1735 if (rdpg.rp_npages > PTLRPC_MAX_BRW_PAGES)
1738 CDEBUG(D_INFO, "reading nodemap log, name '%s', size = %u\n",
1739 body->mcb_name, rdpg.rp_count);
1741 /* allocate pages to store the containers */
1742 OBD_ALLOC_PTR_ARRAY(rdpg.rp_pages, rdpg.rp_npages);
1743 if (rdpg.rp_pages == NULL)
1745 for (i = 0; i < rdpg.rp_npages; i++) {
1746 rdpg.rp_pages[i] = alloc_page(GFP_NOFS);
1747 if (rdpg.rp_pages[i] == NULL)
1748 GOTO(out, rc = -ENOMEM);
1751 rdpg.rp_hash = body->mcb_offset;
1752 nodemap_ii.ii_magic = IDX_INFO_MAGIC;
1753 nodemap_ii.ii_flags = II_FL_NOHASH;
1754 nodemap_ii.ii_version = rqexp_ted->ted_nodemap_version;
1755 nodemap_ii.ii_attrs = body->mcb_nm_cur_pass;
1756 nodemap_ii.ii_count = 0;
1758 bytes = nodemap_index_read(req->rq_svc_thread->t_env,
1759 obd2obt(mgs_obd)->obt_nodemap_config_file,
1760 &nodemap_ii, &rdpg);
1762 GOTO(out, rc = bytes);
1764 rqexp_ted->ted_nodemap_version = nodemap_ii.ii_version;
1766 res = req_capsule_server_get(&req->rq_pill, &RMF_MGS_CONFIG_RES);
1768 GOTO(out, rc = -EINVAL);
1769 res->mcr_offset = nodemap_ii.ii_hash_end;
1770 res->mcr_nm_cur_pass = nodemap_ii.ii_attrs;
1772 page_count = (bytes + PAGE_SIZE - 1) >> PAGE_SHIFT;
1773 LASSERT(page_count <= rdpg.rp_count);
1774 desc = ptlrpc_prep_bulk_exp(req, page_count, 1,
1775 PTLRPC_BULK_PUT_SOURCE,
1776 MGS_BULK_PORTAL, frag_ops);
1778 GOTO(out, rc = -ENOMEM);
1780 for (i = 0; i < page_count && bytes > 0; i++) {
1781 frag_ops->add_kiov_frag(desc, rdpg.rp_pages[i], 0,
1782 min_t(int, bytes, PAGE_SIZE));
1786 rc = target_bulk_io(req->rq_export, desc);
1787 ptlrpc_free_bulk(desc);
1790 if (rdpg.rp_pages != NULL) {
1791 for (i = 0; i < rdpg.rp_npages; i++)
1792 if (rdpg.rp_pages[i] != NULL)
1793 __free_page(rdpg.rp_pages[i]);
1794 OBD_FREE_PTR_ARRAY(rdpg.rp_pages, rdpg.rp_npages);
1798 EXPORT_SYMBOL(nodemap_get_config_req);