4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 only,
8 * as published by the Free Software Foundation.
10 * This program is distributed in the hope that it will be useful, but
11 * WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * General Public License version 2 for more details (a copy is included
14 * in the LICENSE file that accompanied this code).
16 * You should have received a copy of the GNU General Public License
17 * version 2 along with this program; If not, see
18 * http://www.gnu.org/licenses/gpl-2.0.html
23 * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
24 * Use is subject to license terms.
26 * Copyright (c) 2011, 2014, Intel Corporation.
28 * Copyright 2015 Cray Inc, all rights reserved.
31 * Define ost_id associated functions
34 #ifndef _LUSTRE_OSTID_H_
35 #define _LUSTRE_OSTID_H_
37 #include <libcfs/libcfs.h>
38 #include <lustre/lustre_fid.h>
40 static inline __u64 lmm_oi_id(const struct ost_id *oi)
45 static inline __u64 lmm_oi_seq(const struct ost_id *oi)
50 static inline void lmm_oi_set_seq(struct ost_id *oi, __u64 seq)
55 static inline void lmm_oi_set_id(struct ost_id *oi, __u64 oid)
60 static inline void lmm_oi_le_to_cpu(struct ost_id *dst_oi,
61 const struct ost_id *src_oi)
63 dst_oi->oi.oi_id = __le64_to_cpu(src_oi->oi.oi_id);
64 dst_oi->oi.oi_seq = __le64_to_cpu(src_oi->oi.oi_seq);
67 static inline void lmm_oi_cpu_to_le(struct ost_id *dst_oi,
68 const struct ost_id *src_oi)
70 dst_oi->oi.oi_id = __cpu_to_le64(src_oi->oi.oi_id);
71 dst_oi->oi.oi_seq = __cpu_to_le64(src_oi->oi.oi_seq);
74 /* extract OST sequence (group) from a wire ost_id (id/seq) pair */
75 static inline __u64 ostid_seq(const struct ost_id *ostid)
77 if (fid_seq_is_mdt0(ostid->oi.oi_seq))
78 return FID_SEQ_OST_MDT0;
80 if (fid_seq_is_default(ostid->oi.oi_seq))
81 return FID_SEQ_LOV_DEFAULT;
83 if (fid_is_idif(&ostid->oi_fid))
84 return FID_SEQ_OST_MDT0;
86 return fid_seq(&ostid->oi_fid);
89 /* extract OST objid from a wire ost_id (id/seq) pair */
90 static inline __u64 ostid_id(const struct ost_id *ostid)
92 if (fid_seq_is_mdt0(ostid->oi.oi_seq))
93 return ostid->oi.oi_id & IDIF_OID_MASK;
95 if (fid_seq_is_default(ostid->oi.oi_seq))
96 return ostid->oi.oi_id;
98 if (fid_is_idif(&ostid->oi_fid))
99 return fid_idif_id(fid_seq(&ostid->oi_fid),
100 fid_oid(&ostid->oi_fid), 0);
102 return fid_oid(&ostid->oi_fid);
105 static inline void ostid_set_seq(struct ost_id *oi, __u64 seq)
107 if (fid_seq_is_mdt0(seq) || fid_seq_is_default(seq)) {
110 oi->oi_fid.f_seq = seq;
112 * Note: if f_oid + f_ver is zero, we need init it
113 * to be 1, otherwise, ostid_seq will treat this
114 * as old ostid (oi_seq == 0)
116 if (!oi->oi_fid.f_oid && !oi->oi_fid.f_ver)
117 oi->oi_fid.f_oid = LUSTRE_FID_INIT_OID;
121 static inline void ostid_set_seq_mdt0(struct ost_id *oi)
123 ostid_set_seq(oi, FID_SEQ_OST_MDT0);
126 static inline void ostid_set_seq_echo(struct ost_id *oi)
128 ostid_set_seq(oi, FID_SEQ_ECHO);
131 static inline void ostid_set_seq_llog(struct ost_id *oi)
133 ostid_set_seq(oi, FID_SEQ_LLOG);
137 * Note: we need check oi_seq to decide where to set oi_id,
138 * so oi_seq should always be set ahead of oi_id.
140 static inline void ostid_set_id(struct ost_id *oi, __u64 oid)
142 if (fid_seq_is_mdt0(oi->oi.oi_seq)) {
143 if (oid >= IDIF_MAX_OID) {
144 CERROR("Bad %llu to set "DOSTID"\n",
145 (unsigned long long)oid, POSTID(oi));
149 } else if (fid_is_idif(&oi->oi_fid)) {
150 if (oid >= IDIF_MAX_OID) {
151 CERROR("Bad %llu to set "DOSTID"\n",
152 (unsigned long long)oid, POSTID(oi));
155 oi->oi_fid.f_seq = fid_idif_seq(oid,
156 fid_idif_ost_idx(&oi->oi_fid));
157 oi->oi_fid.f_oid = oid;
158 oi->oi_fid.f_ver = oid >> 48;
160 if (oid > OBIF_MAX_OID) {
161 CERROR("Bad %llu to set "DOSTID"\n",
162 (unsigned long long)oid, POSTID(oi));
165 oi->oi_fid.f_oid = oid;
169 static inline void ostid_cpu_to_le(const struct ost_id *src_oi,
170 struct ost_id *dst_oi)
172 if (fid_seq_is_mdt0(src_oi->oi.oi_seq)) {
173 dst_oi->oi.oi_id = __cpu_to_le64(src_oi->oi.oi_id);
174 dst_oi->oi.oi_seq = __cpu_to_le64(src_oi->oi.oi_seq);
176 fid_cpu_to_le(&dst_oi->oi_fid, &src_oi->oi_fid);
180 static inline void ostid_le_to_cpu(const struct ost_id *src_oi,
181 struct ost_id *dst_oi)
183 if (fid_seq_is_mdt0(src_oi->oi.oi_seq)) {
184 dst_oi->oi.oi_id = __le64_to_cpu(src_oi->oi.oi_id);
185 dst_oi->oi.oi_seq = __le64_to_cpu(src_oi->oi.oi_seq);
187 fid_le_to_cpu(&dst_oi->oi_fid, &src_oi->oi_fid);
191 /* pack any OST FID into an ostid (id/seq) for the wire/disk */
192 static inline int fid_to_ostid(const struct lu_fid *fid, struct ost_id *ostid)
194 if (fid_seq_is_igif(fid->f_seq)) {
195 CERROR("bad IGIF, "DFID"\n", PFID(fid));
199 if (fid_is_idif(fid)) {
200 ostid_set_seq_mdt0(ostid);
201 ostid_set_id(ostid, fid_idif_id(fid_seq(fid), fid_oid(fid),
204 ostid->oi_fid = *fid;
211 * Sigh, because pre-2.4 uses
212 * struct lov_mds_md_v1 {
214 * __u64 lmm_object_id;
215 * __u64 lmm_object_seq;
218 * to identify the LOV(MDT) object, and lmm_object_seq will
219 * be normal_fid, which make it hard to combine these conversion
220 * to ostid_to FID. so we will do lmm_oi/fid conversion separately
222 * We can tell the lmm_oi by this way,
223 * 1.8: lmm_object_id = {inode}, lmm_object_gr = 0
224 * 2.1: lmm_object_id = {oid < 128k}, lmm_object_seq = FID_SEQ_NORMAL
225 * 2.4: lmm_oi.f_seq = FID_SEQ_NORMAL, lmm_oi.f_oid = {oid < 128k},
228 * But currently lmm_oi/lsm_oi does not have any "real" usages,
229 * except for printing some information, and the user can always
230 * get the real FID from LMA, besides this multiple case check might
231 * make swab more complicate. So we will keep using id/seq for lmm_oi.
234 static inline void fid_to_lmm_oi(const struct lu_fid *fid,
237 oi->oi.oi_id = fid_oid(fid);
238 oi->oi.oi_seq = fid_seq(fid);
242 * Unpack an OST object id/seq (group) into a FID. This is needed for
243 * converting all obdo, lmm, lsm, etc. 64-bit id/seq pairs into proper
244 * FIDs. Note that if an id/seq is already in FID/IDIF format it will
245 * be passed through unchanged. Only legacy OST objects in "group 0"
246 * will be mapped into the IDIF namespace so that they can fit into the
247 * struct lu_fid fields without loss.
249 static inline int ostid_to_fid(struct lu_fid *fid, const struct ost_id *ostid,
252 __u64 seq = ostid_seq(ostid);
254 if (ost_idx > 0xffff) {
255 CERROR("bad ost_idx, "DOSTID" ost_idx:%u\n", POSTID(ostid),
260 if (fid_seq_is_mdt0(seq)) {
261 __u64 oid = ostid_id(ostid);
263 /* This is a "legacy" (old 1.x/2.early) OST object in "group 0"
264 * that we map into the IDIF namespace. It allows up to 2^48
265 * objects per OST, as this is the object namespace that has
266 * been in production for years. This can handle create rates
267 * of 1M objects/s/OST for 9 years, or combinations thereof.
269 if (oid >= IDIF_MAX_OID) {
270 CERROR("bad MDT0 id(1), "DOSTID" ost_idx:%u\n",
271 POSTID(ostid), ost_idx);
274 fid->f_seq = fid_idif_seq(oid, ost_idx);
275 /* truncate to 32 bits by assignment */
277 /* in theory, not currently used */
278 fid->f_ver = oid >> 48;
279 } else if (!fid_seq_is_default(seq)) {
280 /* This is either an IDIF object, which identifies objects
281 * across all OSTs, or a regular FID. The IDIF namespace
282 * maps legacy OST objects into the FID namespace. In both
283 * cases, we just pass the FID through, no conversion needed.
285 if (ostid->oi_fid.f_ver) {
286 CERROR("bad MDT0 id(2), "DOSTID" ost_idx:%u\n",
287 POSTID(ostid), ost_idx);
290 *fid = ostid->oi_fid;