Whamcloud - gitweb
b=11752
[fs/lustre-release.git] / lustre / lvfs / fsfilt_ext3.c
index a45560a..ad0d2e4 100644 (file)
@@ -25,6 +25,8 @@
 
 #define DEBUG_SUBSYSTEM S_FILTER
 
+#include <linux/init.h>
+#include <linux/module.h>
 #include <linux/fs.h>
 #include <linux/jbd.h>
 #include <linux/slab.h>
 #include <linux/ext3_fs.h>
 #include <linux/ext3_jbd.h>
 #include <linux/version.h>
-/* XXX ugh */
+#include <linux/bitops.h>
+#include <linux/quota.h>
+#include <linux/quotaio_v1.h>
+#include <linux/quotaio_v2.h>
 #if (LINUX_VERSION_CODE < KERNEL_VERSION(2,5,0))
- #include <linux/ext3_xattr.h>
+#include <linux/ext3_xattr.h>
 #else
- #include <linux/../../fs/ext3/xattr.h>
+#include <ext3/xattr.h>
+#endif
+
+#include <libcfs/kp30.h>
+#include <lustre_fsfilt.h>
+#include <obd.h>
+#include <obd_class.h>
+#include <lustre_quota.h>
+#if (LINUX_VERSION_CODE < KERNEL_VERSION(2,5,0))
+#include <linux/iobuf.h>
+#endif
+#include <linux/lustre_compat25.h>
+#include <linux/lprocfs_status.h>
+
+#ifdef EXT3_MULTIBLOCK_ALLOCATOR
+#include <linux/ext3_extents.h>
+#endif
+
+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,15)
+#define FSFILT_DATA_TRANS_BLOCKS(sb)      EXT3_DATA_TRANS_BLOCKS
+#define FSFILT_DELETE_TRANS_BLOCKS(sb)    EXT3_DELETE_TRANS_BLOCKS
+#else
+#define FSFILT_DATA_TRANS_BLOCKS(sb)      EXT3_DATA_TRANS_BLOCKS(sb)
+#define FSFILT_DELETE_TRANS_BLOCKS(sb)    EXT3_DELETE_TRANS_BLOCKS(sb)
 #endif
-#include <linux/kp30.h>
-#include <linux/lustre_fsfilt.h>
-#include <linux/obd.h>
-#include <linux/obd_class.h>
-#include <linux/module.h>
 
 static kmem_cache_t *fcb_cache;
-static atomic_t fcb_cache_count = ATOMIC_INIT(0);
 
 struct fsfilt_cb_data {
         struct journal_callback cb_jcb; /* jbd private data - MUST BE FIRST */
@@ -59,10 +81,49 @@ struct fsfilt_cb_data {
 #ifndef EXT3_XATTR_INDEX_TRUSTED        /* temporary until we hit l28 kernel */
 #define EXT3_XATTR_INDEX_TRUSTED        4
 #endif
-#define XATTR_LUSTRE_MDS_LOV_EA         "lov"
 
-#define EXT3_XATTR_INDEX_LUSTRE         5                         /* old */
-#define XATTR_LUSTRE_MDS_OBJID          "system.lustre_mds_objid" /* old */
+static char *fsfilt_ext3_get_label(struct super_block *sb)
+{
+        return EXT3_SB(sb)->s_es->s_volume_name;
+}
+
+static int fsfilt_ext3_set_label(struct super_block *sb, char *label)
+{
+        /* see e.g. fsfilt_ext3_write_record() */
+        journal_t *journal;
+        handle_t *handle;
+        int err;
+
+        journal = EXT3_SB(sb)->s_journal;
+        lock_24kernel();
+        handle = journal_start(journal, 1);
+        unlock_24kernel();
+        if (IS_ERR(handle)) {
+                CERROR("can't start transaction\n");
+                return(PTR_ERR(handle));
+        }
+
+        err = ext3_journal_get_write_access(handle, EXT3_SB(sb)->s_sbh);
+        if (err)
+                goto out;
+
+        memcpy(EXT3_SB(sb)->s_es->s_volume_name, label,
+               sizeof(EXT3_SB(sb)->s_es->s_volume_name));
+
+        err = ext3_journal_dirty_metadata(handle, EXT3_SB(sb)->s_sbh);
+
+out:
+        lock_24kernel();
+        journal_stop(handle);
+        unlock_24kernel();
+
+        return(err);
+}
+
+static char *fsfilt_ext3_uuid(struct super_block *sb)
+{
+        return EXT3_SB(sb)->s_es->s_uuid;
+}
 
 /*
  * We don't currently need any additional blocks for rmdir and
@@ -70,43 +131,53 @@ struct fsfilt_cb_data {
  * the inode (which we will be changing anyways as part of this
  * transaction).
  */
-static void *fsfilt_ext3_start(struct inode *inode, int op, void *desc_private)
+static void *fsfilt_ext3_start(struct inode *inode, int op, void *desc_private,
+                               int logs)
 {
-        /* For updates to the last recieved file */
-        int nblocks = EXT3_DATA_TRANS_BLOCKS;
-        int blocksize, block_count = 0;
+        /* For updates to the last received file */
+        int nblocks = EXT3_SINGLEDATA_TRANS_BLOCKS;
+        journal_t *journal;
         void *handle;
 
         if (current->journal_info) {
-                CDEBUG(D_INODE, "increasing refcount on %p\n", current->journal_info);
+                CDEBUG(D_INODE, "increasing refcount on %p\n",
+                       current->journal_info);
                 goto journal_start;
         }
 
         switch(op) {
-        case FSFILT_OP_CREATE_LOG:
-                nblocks += EXT3_INDEX_EXTRA_TRANS_BLOCKS+EXT3_DATA_TRANS_BLOCKS;
-                op = FSFILT_OP_CREATE;
-                break;
-        case FSFILT_OP_UNLINK_LOG:
-                nblocks += EXT3_INDEX_EXTRA_TRANS_BLOCKS+EXT3_DATA_TRANS_BLOCKS;
-                op = FSFILT_OP_UNLINK;
-                break;
-        }
-
-        switch(op) {
         case FSFILT_OP_RMDIR:
         case FSFILT_OP_UNLINK:
-                nblocks += EXT3_DELETE_TRANS_BLOCKS;
+                /* delete one file + create/update logs for each stripe */
+                nblocks += FSFILT_DELETE_TRANS_BLOCKS(inode->i_sb);
+                nblocks += (EXT3_INDEX_EXTRA_TRANS_BLOCKS +
+                            EXT3_SINGLEDATA_TRANS_BLOCKS) * logs;
                 break;
         case FSFILT_OP_RENAME:
                 /* modify additional directory */
-                nblocks += EXT3_DATA_TRANS_BLOCKS;
+                nblocks += EXT3_SINGLEDATA_TRANS_BLOCKS;
                 /* no break */
         case FSFILT_OP_SYMLINK:
                 /* additional block + block bitmap + GDT for long symlink */
                 nblocks += 3;
                 /* no break */
-        case FSFILT_OP_CREATE:
+        case FSFILT_OP_CREATE: {
+#if defined(EXT3_EXTENTS_FL) && defined(EXT3_INDEX_FL)
+                static int warned;
+                if (!warned) {
+                        if (!test_opt(inode->i_sb, EXTENTS)) {
+                                warned = 1;
+                        } else if (((EXT3_I(inode)->i_flags &
+                              cpu_to_le32(EXT3_EXTENTS_FL | EXT3_INDEX_FL)) ==
+                              cpu_to_le32(EXT3_EXTENTS_FL | EXT3_INDEX_FL))) {
+                                CWARN("extent-mapped directory found - contact "
+                                      "CFS: support@clusterfs.com\n");
+                                warned = 1;
+                        }
+                }
+#endif
+                /* no break */
+        }
         case FSFILT_OP_MKDIR:
         case FSFILT_OP_MKNOD:
                 /* modify one inode + block bitmap + GDT */
@@ -114,32 +185,63 @@ static void *fsfilt_ext3_start(struct inode *inode, int op, void *desc_private)
                 /* no break */
         case FSFILT_OP_LINK:
                 /* modify parent directory */
-                nblocks += EXT3_INDEX_EXTRA_TRANS_BLOCKS+EXT3_DATA_TRANS_BLOCKS;
+                nblocks += EXT3_INDEX_EXTRA_TRANS_BLOCKS +
+                         FSFILT_DATA_TRANS_BLOCKS(inode->i_sb);
+                /* create/update logs for each stripe */
+                nblocks += (EXT3_INDEX_EXTRA_TRANS_BLOCKS +
+                            EXT3_SINGLEDATA_TRANS_BLOCKS) * logs;
                 break;
         case FSFILT_OP_SETATTR:
                 /* Setattr on inode */
                 nblocks += 1;
+                nblocks += EXT3_INDEX_EXTRA_TRANS_BLOCKS +
+                         FSFILT_DATA_TRANS_BLOCKS(inode->i_sb);
+                /* quota chown log for each stripe */
+                nblocks += (EXT3_INDEX_EXTRA_TRANS_BLOCKS +
+                            EXT3_SINGLEDATA_TRANS_BLOCKS) * logs;
                 break;
-        case FSFILT_OP_CANCEL_UNLINK_LOG:
-                blocksize = 1 << inode->i_blkbits;
-                block_count = (blocksize - 1) + LLOG_CHUNK_SIZE;
-                block_count = (block_count + blocksize - 1) >> inode->i_blkbits;
-                block_count = block_count * EXT3_DATA_TRANS_BLOCKS + 2;
-                nblocks = 2 * 2 * block_count;
+        case FSFILT_OP_CANCEL_UNLINK:
+                /* blocks for log header bitmap update OR
+                 * blocks for catalog header bitmap update + unlink of logs */
+                nblocks = (LLOG_CHUNK_SIZE >> inode->i_blkbits) +
+                        FSFILT_DELETE_TRANS_BLOCKS(inode->i_sb) * logs;
+                break;
+        case FSFILT_OP_JOIN:
+                /* delete 2 file(file + array id) + create 1 file (array id) 
+                 * create/update logs for each stripe */
+                nblocks += 2 * FSFILT_DELETE_TRANS_BLOCKS(inode->i_sb);
+               
+                /*create array log for head file*/ 
+                nblocks += 3;
+                nblocks += (EXT3_INDEX_EXTRA_TRANS_BLOCKS +
+                            EXT3_SINGLEDATA_TRANS_BLOCKS);
+                /*update head file array */
+                nblocks += EXT3_INDEX_EXTRA_TRANS_BLOCKS +
+                         FSFILT_DATA_TRANS_BLOCKS(inode->i_sb);
                 break;
         default: CERROR("unknown transaction start op %d\n", op);
-                 LBUG();
+                LBUG();
         }
 
         LASSERT(current->journal_info == desc_private);
+        journal = EXT3_SB(inode->i_sb)->s_journal;
+        if (nblocks > journal->j_max_transaction_buffers) {
+                CWARN("too many credits %d for op %ux%u using %d instead\n",
+                      nblocks, op, logs, journal->j_max_transaction_buffers);
+                nblocks = journal->j_max_transaction_buffers;
+        }
 
  journal_start:
-        lock_kernel();
+        LASSERTF(nblocks > 0, "can't start %d credit transaction\n", nblocks);
+        lock_24kernel();
         handle = journal_start(EXT3_JOURNAL(inode), nblocks);
-        unlock_kernel();
+        unlock_24kernel();
 
         if (!IS_ERR(handle))
                 LASSERT(current->journal_info == handle);
+        else
+                CERROR("error starting handle for op %u (%u credits): rc %ld\n",
+                       op, nblocks, PTR_ERR(handle));
         return handle;
 }
 
@@ -172,7 +274,7 @@ static int fsfilt_ext3_credits_needed(int objcount, struct fsfilt_objinfo *fso,
 {
         struct super_block *sb = fso->fso_dentry->d_inode->i_sb;
         __u64 next_indir;
-        const int blockpp = 1 << (PAGE_CACHE_SHIFT - sb->s_blocksize_bits);
+        const int blockpp = 1 << (CFS_PAGE_SHIFT - sb->s_blocksize_bits);
         int nbitmaps = 0, ngdblocks;
         int needed = objcount + 1; /* inodes + superblock */
         int i, j;
@@ -213,9 +315,9 @@ static int fsfilt_ext3_credits_needed(int objcount, struct fsfilt_objinfo *fso,
         needed += nbitmaps + ngdblocks;
 
         /* last_rcvd update */
-        needed += EXT3_DATA_TRANS_BLOCKS;
+        needed += FSFILT_DATA_TRANS_BLOCKS(sb);
 
-#if defined(CONFIG_QUOTA) && !defined(__x86_64__) /* XXX */
+#if defined(CONFIG_QUOTA)
         /* We assume that there will be 1 bit set in s_dquot.flags for each
          * quota file that is active.  This is at least true for now.
          */
@@ -242,7 +344,7 @@ static int fsfilt_ext3_credits_needed(int objcount, struct fsfilt_objinfo *fso,
  */
 static void *fsfilt_ext3_brw_start(int objcount, struct fsfilt_objinfo *fso,
                                    int niocount, struct niobuf_local *nb,
-                                   void *desc_private)
+                                   void *desc_private, int logs)
 {
         journal_t *journal;
         handle_t *handle;
@@ -269,9 +371,10 @@ static void *fsfilt_ext3_brw_start(int objcount, struct fsfilt_objinfo *fso,
                 needed = journal->j_max_transaction_buffers;
         }
 
-        lock_kernel();
+        LASSERTF(needed > 0, "can't start %d credit transaction\n", needed);
+        lock_24kernel();
         handle = journal_start(journal, needed);
-        unlock_kernel();
+        unlock_24kernel();
         if (IS_ERR(handle)) {
                 CERROR("can't get handle for %d credits: rc = %ld\n", needed,
                        PTR_ERR(handle));
@@ -283,6 +386,25 @@ static void *fsfilt_ext3_brw_start(int objcount, struct fsfilt_objinfo *fso,
         RETURN(handle);
 }
 
+static int fsfilt_ext3_extend(struct inode *inode, unsigned int nblocks,void *h)
+{
+       handle_t *handle = h;
+
+       /* fsfilt_extend called with nblocks = 0 for testing in special cases */
+       if (nblocks == 0) {
+               handle->h_buffer_credits = 0;
+               CWARN("setting credits of handle %p to zero by request\n", h);
+       }
+
+       if (handle->h_buffer_credits > nblocks)
+                return 0;
+       if (journal_extend(handle, nblocks) == 0)
+                return 0;
+
+       ext3_mark_inode_dirty(handle, inode);
+       return journal_restart(handle, nblocks);
+}
+
 static int fsfilt_ext3_commit(struct inode *inode, void *h, int force_sync)
 {
         int rc;
@@ -292,26 +414,28 @@ static int fsfilt_ext3_commit(struct inode *inode, void *h, int force_sync)
         if (force_sync)
                 handle->h_sync = 1; /* recovery likes this */
 
-        lock_kernel();
+        lock_24kernel();
         rc = journal_stop(handle);
-        unlock_kernel();
+        unlock_24kernel();
 
-        // LASSERT(current->journal_info == NULL);
         return rc;
 }
 
 static int fsfilt_ext3_commit_async(struct inode *inode, void *h,
-                                        void **wait_handle)
+                                    void **wait_handle)
 {
+        unsigned long tid;
         transaction_t *transaction;
-        unsigned long tid, rtid;
+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,5,0)
+        unsigned long rtid;
+#endif
         handle_t *handle = h;
         journal_t *journal;
         int rc;
 
         LASSERT(current->journal_info == handle);
 
-        lock_kernel();
+        lock_24kernel();
         transaction = handle->h_transaction;
         journal = transaction->t_journal;
         tid = transaction->t_tid;
@@ -320,7 +444,7 @@ static int fsfilt_ext3_commit_async(struct inode *inode, void *h,
         rc = journal_stop(handle);
         if (rc) {
                 CERROR("error while stopping transaction: %d\n", rc);
-                unlock_kernel();
+                unlock_24kernel();
                 return rc;
         }
 #if LINUX_VERSION_CODE < KERNEL_VERSION(2,5,0)
@@ -329,9 +453,9 @@ static int fsfilt_ext3_commit_async(struct inode *inode, void *h,
                 CERROR("strange race: %lu != %lu\n",
                        (unsigned long) tid, (unsigned long) rtid);
 #else
-        log_start_commit(journal, transaction->t_tid);
+        log_start_commit(journal, tid);
 #endif
-        unlock_kernel();
+        unlock_24kernel();
 
         *wait_handle = (void *) tid;
         CDEBUG(D_INODE, "commit async: %lu\n", (unsigned long) tid);
@@ -340,14 +464,17 @@ static int fsfilt_ext3_commit_async(struct inode *inode, void *h,
 
 static int fsfilt_ext3_commit_wait(struct inode *inode, void *h)
 {
+        journal_t *journal = EXT3_JOURNAL(inode);
         tid_t tid = (tid_t)(long)h;
 
         CDEBUG(D_INODE, "commit wait: %lu\n", (unsigned long) tid);
-       if (is_journal_aborted(EXT3_JOURNAL(inode)))
+        if (unlikely(is_journal_aborted(journal)))
                 return -EIO;
 
         log_wait_commit(EXT3_JOURNAL(inode), tid);
 
+        if (unlikely(is_journal_aborted(journal)))
+                return -EIO;
         return 0;
 }
 
@@ -355,30 +482,45 @@ static int fsfilt_ext3_setattr(struct dentry *dentry, void *handle,
                                struct iattr *iattr, int do_trunc)
 {
         struct inode *inode = dentry->d_inode;
-        int rc;
+        int rc = 0;
 
         lock_kernel();
 
-        /* A _really_ horrible hack to avoid removing the data stored
-         * in the block pointers; this is really the "small" stripe MD data.
-         * We can avoid further hackery by virtue of the MDS file size being
-         * zero all the time (which doesn't invoke block truncate at unlink
-         * time), so we assert we never change the MDS file size from zero. */
+        /* Avoid marking the inode dirty on the superblock list unnecessarily.
+         * We are already writing the inode to disk as part of this
+         * transaction and want to avoid a lot of extra inode writeout
+         * later on. b=9828 */
         if (iattr->ia_valid & ATTR_SIZE && !do_trunc) {
                 /* ATTR_SIZE would invoke truncate: clear it */
                 iattr->ia_valid &= ~ATTR_SIZE;
                 EXT3_I(inode)->i_disksize = inode->i_size = iattr->ia_size;
 
-                /* make sure _something_ gets set - so new inode
-                 * goes to disk (probably won't work over XFS */
-                if (!(iattr->ia_valid & (ATTR_MODE | ATTR_MTIME | ATTR_CTIME))){
-                        iattr->ia_valid |= ATTR_MODE;
-                        iattr->ia_mode = inode->i_mode;
+                if (iattr->ia_valid & ATTR_UID)
+                        inode->i_uid = iattr->ia_uid;
+                if (iattr->ia_valid & ATTR_GID)
+                        inode->i_gid = iattr->ia_gid;
+                if (iattr->ia_valid & ATTR_ATIME)
+                        inode->i_atime = iattr->ia_atime;
+                if (iattr->ia_valid & ATTR_MTIME)
+                        inode->i_mtime = iattr->ia_mtime;
+                if (iattr->ia_valid & ATTR_CTIME)
+                        inode->i_ctime = iattr->ia_ctime;
+                if (iattr->ia_valid & ATTR_MODE) {
+                        inode->i_mode = iattr->ia_mode;
+
+                        if (!in_group_p(inode->i_gid) && !capable(CAP_FSETID))
+                                inode->i_mode &= ~S_ISGID;
                 }
+
+                inode->i_sb->s_op->dirty_inode(inode);
+
+                goto out;
         }
 
         /* Don't allow setattr to change file type */
-        iattr->ia_mode = (inode->i_mode & S_IFMT)|(iattr->ia_mode & ~S_IFMT);
+        if (iattr->ia_valid & ATTR_MODE)
+                iattr->ia_mode = (inode->i_mode & S_IFMT) |
+                                 (iattr->ia_mode & ~S_IFMT);
 
         /* We set these flags on the client, but have already checked perms
          * so don't confuse inode_change_ok. */
@@ -392,9 +534,9 @@ static int fsfilt_ext3_setattr(struct dentry *dentry, void *handle,
                         rc = inode_setattr(inode, iattr);
         }
 
+ out:
         unlock_kernel();
-
-        return rc;
+        RETURN(rc);
 }
 
 static int fsfilt_ext3_iocontrol(struct inode * inode, struct file *file,
@@ -403,6 +545,12 @@ static int fsfilt_ext3_iocontrol(struct inode * inode, struct file *file,
         int rc = 0;
         ENTRY;
 
+        /* FIXME: Can't do this because of nested transaction deadlock */
+        if (cmd == EXT3_IOC_SETFLAGS && (*(int *)arg) & EXT3_JOURNAL_DATA_FL) {
+                CERROR("can't set data journal flag on file\n");
+                RETURN(-EPERM);
+        }
+
         if (inode->i_fop->ioctl)
                 rc = inode->i_fop->ioctl(inode, file, cmd, arg);
         else
@@ -411,72 +559,18 @@ static int fsfilt_ext3_iocontrol(struct inode * inode, struct file *file,
         RETURN(rc);
 }
 
-#undef INLINE_EA
-#undef OLD_EA
 static int fsfilt_ext3_set_md(struct inode *inode, void *handle,
-                              void *lmm, int lmm_size)
+                              void *lmm, int lmm_size, const char *name)
 {
-        int rc, old_ea = 0;
-
-#ifdef INLINE_EA  /* can go away before 1.0 - just for testing bug 2097 now */
-        /* Nasty hack city - store stripe MD data in the block pointers if
-         * it will fit, because putting it in an EA currently kills the MDS
-         * performance.  We'll fix this with "fast EAs" in the future.
-         */
-        if (inode->i_blocks == 0 && lmm_size <= sizeof(EXT3_I(inode)->i_data) -
-                                            sizeof(EXT3_I(inode)->i_data[0])) {
-                unsigned old_size = EXT3_I(inode)->i_data[0];
-                if (old_size != 0) {
-                        LASSERT(old_size < sizeof(EXT3_I(inode)->i_data));
-                        CERROR("setting EA on %lu/%u again... interesting\n",
-                               inode->i_ino, inode->i_generation);
-                }
+        int rc;
 
-                EXT3_I(inode)->i_data[0] = cpu_to_le32(lmm_size);
-                memcpy(&EXT3_I(inode)->i_data[1], lmm, lmm_size);
-                mark_inode_dirty(inode);
-                return 0;
-        }
-#endif
-#ifdef OLD_EA
-        /* keep this when we get rid of OLD_EA (too noisy during conversion) */
-        if (EXT3_I(inode)->i_file_acl /* || large inode EA flag */) {
-                CWARN("setting EA on %lu/%u again... interesting\n",
-                       inode->i_ino, inode->i_generation);
-                old_ea = 1;
-        }
+        LASSERT(TRYLOCK_INODE_MUTEX(inode) == 0);
 
-        lock_kernel();
-        /* this can go away before 1.0.  For bug 2097 testing only. */
-        rc = ext3_xattr_set_handle(handle, inode, EXT3_XATTR_INDEX_LUSTRE,
-                                   XATTR_LUSTRE_MDS_OBJID, lmm, lmm_size, 0);
-#else
-        lock_kernel();
+        lock_24kernel();
         rc = ext3_xattr_set_handle(handle, inode, EXT3_XATTR_INDEX_TRUSTED,
-                                   XATTR_LUSTRE_MDS_LOV_EA, lmm, lmm_size, 0);
-
-        /* This tries to delete the old-format LOV EA, but only as long as we
-         * have successfully saved the new-format LOV EA (we can always try
-         * the conversion again the next time the file is accessed).  It is
-         * possible (although unlikely) that the new-format LOV EA couldn't be
-         * saved because it ran out of space but we would need a file striped
-         * over least 123 OSTs before the two EAs filled a 4kB block.
-         *
-         * This can be removed when all filesystems have converted to the
-         * new EA format, but otherwise adds little if any overhead.  If we
-         * wanted backward compatibility for existing files, we could keep
-         * the old EA around for a while but we'd have to clean it up later. */
-        if (rc >= 0 && old_ea) {
-                int err = ext3_xattr_set_handle(handle, inode,
-                                                EXT3_XATTR_INDEX_LUSTRE,
-                                                XATTR_LUSTRE_MDS_OBJID,
-                                                NULL, 0, 0);
-                if (err)
-                        CERROR("error deleting old LOV EA on %lu/%u: rc %d\n",
-                               inode->i_ino, inode->i_generation, err);
-        }
-#endif
-        unlock_kernel();
+                                   name, lmm, lmm_size, 0);
+
+        unlock_24kernel();
 
         if (rc)
                 CERROR("error adding MD data to inode %lu: rc = %d\n",
@@ -484,69 +578,18 @@ static int fsfilt_ext3_set_md(struct inode *inode, void *handle,
         return rc;
 }
 
-/* Must be called with i_sem held */
-static int fsfilt_ext3_get_md(struct inode *inode, void *lmm, int lmm_size)
+/* Must be called with i_mutex held */
+static int fsfilt_ext3_get_md(struct inode *inode, void *lmm, int lmm_size,
+                              const char *name)
 {
         int rc;
 
-        LASSERT(down_trylock(&inode->i_sem) != 0);
-        lock_kernel();
-        /* Keep support for reading "inline EAs" until we convert
-         * users over to new format entirely.  See bug 841/2097. */
-        if (inode->i_blocks == 0 && EXT3_I(inode)->i_data[0]) {
-                unsigned size = le32_to_cpu(EXT3_I(inode)->i_data[0]);
-                void *handle;
-
-                LASSERT(size < sizeof(EXT3_I(inode)->i_data));
-                if (lmm) {
-                        if (size > lmm_size) {
-                                CERROR("inline EA on %lu/%u bad size %u > %u\n",
-                                       inode->i_ino, inode->i_generation,
-                                       size, lmm_size);
-                                return -ERANGE;
-                        }
-                        memcpy(lmm, &EXT3_I(inode)->i_data[1], size);
-                }
-
-#ifndef INLINE_EA
-                /* migrate LOV EA data to external block - keep same format */
-                CWARN("DEBUG: migrate inline EA for inode %lu/%u to block\n",
-                      inode->i_ino, inode->i_generation);
-
-                handle = journal_start(EXT3_JOURNAL(inode),
-                                       EXT3_XATTR_TRANS_BLOCKS);
-                if (!IS_ERR(handle)) {
-                        int err;
-                        rc = fsfilt_ext3_set_md(inode, handle,
-                                                &EXT3_I(inode)->i_data[1],size);
-                        if (rc == 0) {
-                                memset(EXT3_I(inode)->i_data, 0,
-                                       sizeof(EXT3_I(inode)->i_data));
-                                mark_inode_dirty(inode);
-                        }
-                        err = journal_stop(handle);
-                        if (err && rc == 0)
-                                rc = err;
-                } else {
-                        rc = PTR_ERR(handle);
-                }
-#endif
-                unlock_kernel();
-                return size;
-        }
+        LASSERT(TRYLOCK_INODE_MUTEX(inode) == 0);
+        lock_24kernel();
 
         rc = ext3_xattr_get(inode, EXT3_XATTR_INDEX_TRUSTED,
-                            XATTR_LUSTRE_MDS_LOV_EA, lmm, lmm_size);
-        /* try old EA type if new one failed - MDS will convert it for us */
-        if (rc == -ENODATA) {
-                CDEBUG(D_INFO,"failed new LOV EA %d/%s from inode %lu: rc %d\n",
-                       EXT3_XATTR_INDEX_TRUSTED, XATTR_LUSTRE_MDS_LOV_EA,
-                       inode->i_ino, rc);
-
-                rc = ext3_xattr_get(inode, EXT3_XATTR_INDEX_LUSTRE,
-                                    XATTR_LUSTRE_MDS_OBJID, lmm, lmm_size);
-        }
-        unlock_kernel();
+                            name, lmm, lmm_size);
+        unlock_24kernel();
 
         /* This gives us the MD size */
         if (lmm == NULL)
@@ -554,7 +597,7 @@ static int fsfilt_ext3_get_md(struct inode *inode, void *lmm, int lmm_size)
 
         if (rc < 0) {
                 CDEBUG(D_INFO, "error getting EA %d/%s from inode %lu: rc %d\n",
-                       EXT3_XATTR_INDEX_LUSTRE, XATTR_LUSTRE_MDS_OBJID,
+                       EXT3_XATTR_INDEX_TRUSTED, name,
                        inode->i_ino, rc);
                 memset(lmm, 0, lmm_size);
                 return (rc == -ENODATA) ? 0 : rc;
@@ -563,6 +606,44 @@ static int fsfilt_ext3_get_md(struct inode *inode, void *lmm, int lmm_size)
         return rc;
 }
 
+#if (LINUX_VERSION_CODE > KERNEL_VERSION(2,5,0))
+static int fsfilt_ext3_send_bio(int rw, struct inode *inode, struct bio *bio)
+{
+        submit_bio(rw, bio);
+        return 0;
+}
+#else
+static int fsfilt_ext3_send_bio(int rw, struct inode *inode, struct kiobuf *bio)
+{
+        int rc, blk_per_page;
+
+        rc = brw_kiovec(rw, 1, &bio, inode->i_dev,
+                        KIOBUF_GET_BLOCKS(bio), 1 << inode->i_blkbits);
+        /*
+         * brw_kiovec() returns number of bytes actually written. If error
+         * occurred after something was written, error code is returned though
+         * kiobuf->errno. (See bug 6854.)
+         */
+
+        blk_per_page = CFS_PAGE_SIZE >> inode->i_blkbits;
+
+        if (rc != (1 << inode->i_blkbits) * bio->nr_pages * blk_per_page) {
+                CERROR("short write?  expected %d, wrote %d (%d)\n",
+                       (1 << inode->i_blkbits) * bio->nr_pages * blk_per_page,
+                       rc, bio->errno);
+        }
+        if (bio->errno != 0) {
+                CERROR("IO error. Wrote %d of %d (%d)\n",
+                       rc,
+                       (1 << inode->i_blkbits) * bio->nr_pages * blk_per_page,
+                       bio->errno);
+                rc = bio->errno;
+        }
+
+        return rc;
+}
+#endif
+
 static ssize_t fsfilt_ext3_readpage(struct file *file, char *buf, size_t count,
                                     loff_t *off)
 {
@@ -607,7 +688,7 @@ static ssize_t fsfilt_ext3_readpage(struct file *file, char *buf, size_t count,
 
                                 CDEBUG(D_EXT2, "fake %u@%llu\n", blksize, *off);
                                 memset(fake, 0, sizeof(*fake));
-                                fake->rec_len = cpu_to_le32(blksize);
+                                fake->rec_len = cpu_to_le16(blksize);
                         }
                         count -= blksize;
                         buf += blksize;
@@ -626,7 +707,6 @@ static void fsfilt_ext3_cb_func(struct journal_callback *jcb, int error)
         fcb->cb_func(fcb->cb_obd, fcb->cb_last_rcvd, fcb->cb_data, error);
 
         OBD_SLAB_FREE(fcb, fcb_cache, sizeof *fcb);
-        atomic_dec(&fcb_cache_count);
 }
 
 static int fsfilt_ext3_add_journal_cb(struct obd_device *obd, __u64 last_rcvd,
@@ -639,17 +719,16 @@ static int fsfilt_ext3_add_journal_cb(struct obd_device *obd, __u64 last_rcvd,
         if (fcb == NULL)
                 RETURN(-ENOMEM);
 
-        atomic_inc(&fcb_cache_count);
         fcb->cb_func = cb_func;
         fcb->cb_obd = obd;
         fcb->cb_last_rcvd = last_rcvd;
         fcb->cb_data = cb_data;
 
         CDEBUG(D_EXT2, "set callback for last_rcvd: "LPD64"\n", last_rcvd);
-        lock_kernel();
+        lock_24kernel();
         journal_callback_set(handle, fsfilt_ext3_cb_func,
                              (struct journal_callback *)fcb);
-        unlock_kernel();
+        unlock_24kernel();
 
         return 0;
 }
@@ -664,7 +743,10 @@ static int fsfilt_ext3_add_journal_cb(struct obd_device *obd, __u64 last_rcvd,
 static int fsfilt_ext3_statfs(struct super_block *sb, struct obd_statfs *osfs)
 {
         struct kstatfs sfs;
-        int rc = vfs_statfs(sb, &sfs);
+        int rc;
+
+        memset(&sfs, 0, sizeof(sfs));
+        rc = ll_do_statfs(sb,&sfs);
 
         if (!rc && sfs.f_bfree < sfs.f_ffree) {
                 sfs.f_files = (sfs.f_files - sfs.f_ffree) + sfs.f_bfree;
@@ -680,12 +762,360 @@ static int fsfilt_ext3_sync(struct super_block *sb)
         return ext3_force_commit(sb);
 }
 
+#if defined(EXT3_MULTIBLOCK_ALLOCATOR) && (!defined(EXT3_EXT_CACHE_NO) || defined(EXT_CACHE_MARK))
+#warning "kernel code has old extents/mballoc patch, disabling"
+#undef EXT3_MULTIBLOCK_ALLOCATOR
+#endif
+#ifndef EXT3_EXTENTS_FL
+#define EXT3_EXTENTS_FL                 0x00080000 /* Inode uses extents */
+#endif
+
+#ifdef EXT3_MULTIBLOCK_ALLOCATOR
+#if (LINUX_VERSION_CODE < KERNEL_VERSION(2,5,0))
+#define ext3_up_truncate_sem(inode)  up_write(&EXT3_I(inode)->truncate_sem);
+#define ext3_down_truncate_sem(inode)  down_write(&EXT3_I(inode)->truncate_sem);
+#elif (LINUX_VERSION_CODE < KERNEL_VERSION(2,6,17))
+#define ext3_up_truncate_sem(inode)  up(&EXT3_I(inode)->truncate_sem);
+#define ext3_down_truncate_sem(inode)  down(&EXT3_I(inode)->truncate_sem);
+#else
+#define ext3_up_truncate_sem(inode)  mutex_unlock(&EXT3_I(inode)->truncate_mutex);
+#define ext3_down_truncate_sem(inode)  mutex_lock(&EXT3_I(inode)->truncate_mutex);
+#endif
+
+#include <linux/lustre_version.h>
+#if EXT3_EXT_MAGIC == 0xf301
+#define ee_start e_start
+#define ee_block e_block
+#define ee_len   e_num
+#endif
+#ifndef EXT3_BB_MAX_BLOCKS
+#define ext3_mb_new_blocks(handle, inode, goal, count, aflags, err) \
+        ext3_new_blocks(handle, inode, count, goal, err)
+#endif
+
+struct bpointers {
+        unsigned long *blocks;
+        int *created;
+        unsigned long start;
+        int num;
+        int init_num;
+        int create;
+};
+
+static int ext3_ext_find_goal(struct inode *inode, struct ext3_ext_path *path,
+                              unsigned long block, int *aflags)
+{
+        struct ext3_inode_info *ei = EXT3_I(inode);
+        unsigned long bg_start;
+        unsigned long colour;
+        int depth;
+
+        if (path) {
+                struct ext3_extent *ex;
+                depth = path->p_depth;
+
+                /* try to predict block placement */
+                if ((ex = path[depth].p_ext)) {
+#if 0
+                        /* This prefers to eat into a contiguous extent
+                         * rather than find an extent that the whole
+                         * request will fit into.  This can fragment data
+                         * block allocation and prevents our lovely 1M I/Os
+                         * from reaching the disk intact. */
+                        if (ex->ee_block + ex->ee_len == block)
+                                *aflags |= 1;
+#endif
+                        return ex->ee_start + (block - ex->ee_block);
+                }
+
+                /* it looks index is empty
+                 * try to find starting from index itself */
+                if (path[depth].p_bh)
+                        return path[depth].p_bh->b_blocknr;
+        }
+
+        /* OK. use inode's group */
+        bg_start = (ei->i_block_group * EXT3_BLOCKS_PER_GROUP(inode->i_sb)) +
+                le32_to_cpu(EXT3_SB(inode->i_sb)->s_es->s_first_data_block);
+        colour = (current->pid % 16) *
+                (EXT3_BLOCKS_PER_GROUP(inode->i_sb) / 16);
+        return bg_start + colour + block;
+}
+
+#if (LINUX_VERSION_CODE < KERNEL_VERSION(2,5,0))
+#include <linux/locks.h>
+static void ll_unmap_underlying_metadata(struct super_block *sb,
+                                         unsigned long blocknr)
+{
+        struct buffer_head *old_bh;
+
+        old_bh = get_hash_table(sb->s_dev, blocknr, sb->s_blocksize);
+        if (old_bh) {
+                mark_buffer_clean(old_bh);
+                wait_on_buffer(old_bh);
+                clear_bit(BH_Req, &old_bh->b_state);
+                __brelse(old_bh);
+        }
+}
+#else
+#define ll_unmap_underlying_metadata(sb, blocknr) \
+        unmap_underlying_metadata((sb)->s_bdev, blocknr)
+#endif
+
+static int ext3_ext_new_extent_cb(struct ext3_extents_tree *tree,
+                                  struct ext3_ext_path *path,
+                                  struct ext3_ext_cache *cex)
+{
+        struct inode *inode = tree->inode;
+        struct bpointers *bp = tree->private;
+        struct ext3_extent nex;
+        int count, err, goal;
+        unsigned long pblock;
+        unsigned long tgen;
+        handle_t *handle;
+        int i, aflags = 0;
+
+        i = EXT_DEPTH(tree);
+        EXT_ASSERT(i == path->p_depth);
+        EXT_ASSERT(path[i].p_hdr);
+
+        if (cex->ec_type == EXT3_EXT_CACHE_EXTENT) {
+                err = EXT_CONTINUE;
+                goto map;
+        }
+
+        if (bp->create == 0) {
+                i = 0;
+                if (cex->ec_block < bp->start)
+                        i = bp->start - cex->ec_block;
+                if (i >= cex->ec_len)
+                        CERROR("nothing to do?! i = %d, e_num = %u\n",
+                                        i, cex->ec_len);
+                for (; i < cex->ec_len && bp->num; i++) {
+                        *(bp->created) = 0;
+                        bp->created++;
+                        *(bp->blocks) = 0;
+                        bp->blocks++;
+                        bp->num--;
+                        bp->start++;
+                }
+
+                return EXT_CONTINUE;
+        }
+
+        tgen = EXT_GENERATION(tree);
+        count = ext3_ext_calc_credits_for_insert(tree, path);
+        ext3_up_truncate_sem(inode);
+
+        lock_24kernel();
+        handle = journal_start(EXT3_JOURNAL(inode), count+EXT3_ALLOC_NEEDED+1);
+        unlock_24kernel();
+        if (IS_ERR(handle)) {
+                ext3_down_truncate_sem(inode);
+                return PTR_ERR(handle);
+        }
+
+        ext3_down_truncate_sem(inode);
+        if (tgen != EXT_GENERATION(tree)) {
+                /* the tree has changed. so path can be invalid at moment */
+                lock_24kernel();
+                journal_stop(handle);
+                unlock_24kernel();
+                return EXT_REPEAT;
+        }
+
+        count = cex->ec_len;
+        goal = ext3_ext_find_goal(inode, path, cex->ec_block, &aflags);
+        aflags |= 2; /* block have been already reserved */
+        lock_24kernel();
+        pblock = ext3_mb_new_blocks(handle, inode, goal, &count, aflags, &err);
+        unlock_24kernel();
+        if (!pblock)
+                goto out;
+        EXT_ASSERT(count <= cex->ec_len);
+
+        /* insert new extent */
+        nex.ee_block = cex->ec_block;
+        nex.ee_start = pblock;
+        nex.ee_len = count;
+        err = ext3_ext_insert_extent(handle, tree, path, &nex);
+        if (err)
+                goto out;
+
+        /*
+         * Putting len of the actual extent we just inserted,
+         * we are asking ext3_ext_walk_space() to continue
+         * scaning after that block
+         */
+        cex->ec_len = nex.ee_len;
+        cex->ec_start = nex.ee_start;
+        BUG_ON(nex.ee_len == 0);
+        BUG_ON(nex.ee_block != cex->ec_block);
+
+out:
+        lock_24kernel();
+        journal_stop(handle);
+        unlock_24kernel();
+map:
+        if (err >= 0) {
+                /* map blocks */
+                if (bp->num == 0) {
+                        CERROR("hmm. why do we find this extent?\n");
+                        CERROR("initial space: %lu:%u\n",
+                                bp->start, bp->init_num);
+                        CERROR("current extent: %u/%u/%u %d\n",
+                                cex->ec_block, cex->ec_len,
+                                cex->ec_start, cex->ec_type);
+                }
+                i = 0;
+                if (cex->ec_block < bp->start)
+                        i = bp->start - cex->ec_block;
+                if (i >= cex->ec_len)
+                        CERROR("nothing to do?! i = %d, e_num = %u\n",
+                                        i, cex->ec_len);
+                for (; i < cex->ec_len && bp->num; i++) {
+                        *(bp->blocks) = cex->ec_start + i;
+                        if (cex->ec_type == EXT3_EXT_CACHE_EXTENT) {
+                                *(bp->created) = 0;
+                        } else {
+                                *(bp->created) = 1;
+                                /* unmap any possible underlying metadata from
+                                 * the block device mapping.  bug 6998. */
+                                ll_unmap_underlying_metadata(inode->i_sb,
+                                                             *(bp->blocks));
+                        }
+                        bp->created++;
+                        bp->blocks++;
+                        bp->num--;
+                        bp->start++;
+                }
+        }
+        return err;
+}
+
+int fsfilt_map_nblocks(struct inode *inode, unsigned long block,
+                       unsigned long num, unsigned long *blocks,
+                       int *created, int create)
+{
+        struct ext3_extents_tree tree;
+        struct bpointers bp;
+        int err;
+
+        CDEBUG(D_OTHER, "blocks %lu-%lu requested for inode %u\n",
+               block, block + num - 1, (unsigned) inode->i_ino);
+
+        ext3_init_tree_desc(&tree, inode);
+        tree.private = &bp;
+        bp.blocks = blocks;
+        bp.created = created;
+        bp.start = block;
+        bp.init_num = bp.num = num;
+        bp.create = create;
+
+        ext3_down_truncate_sem(inode);
+        err = ext3_ext_walk_space(&tree, block, num, ext3_ext_new_extent_cb);
+        ext3_ext_invalidate_cache(&tree);
+        ext3_up_truncate_sem(inode);
+
+        return err;
+}
+
+int fsfilt_ext3_map_ext_inode_pages(struct inode *inode, struct page **page,
+                                    int pages, unsigned long *blocks,
+                                    int *created, int create)
+{
+        int blocks_per_page = CFS_PAGE_SIZE >> inode->i_blkbits;
+        int rc = 0, i = 0;
+        struct page *fp = NULL;
+        int clen = 0;
+
+        CDEBUG(D_OTHER, "inode %lu: map %d pages from %lu\n",
+                inode->i_ino, pages, (*page)->index);
+
+        /* pages are sorted already. so, we just have to find
+         * contig. space and process them properly */
+        while (i < pages) {
+                if (fp == NULL) {
+                        /* start new extent */
+                        fp = *page++;
+                        clen = 1;
+                        i++;
+                        continue;
+                } else if (fp->index + clen == (*page)->index) {
+                        /* continue the extent */
+                        page++;
+                        clen++;
+                        i++;
+                        continue;
+                }
+
+                /* process found extent */
+                rc = fsfilt_map_nblocks(inode, fp->index * blocks_per_page,
+                                        clen * blocks_per_page, blocks,
+                                        created, create);
+                if (rc)
+                        GOTO(cleanup, rc);
+
+                /* look for next extent */
+                fp = NULL;
+                blocks += blocks_per_page * clen;
+                created += blocks_per_page * clen;
+        }
+
+        if (fp)
+                rc = fsfilt_map_nblocks(inode, fp->index * blocks_per_page,
+                                        clen * blocks_per_page, blocks,
+                                        created, create);
+cleanup:
+        return rc;
+}
+#endif /* EXT3_MULTIBLOCK_ALLOCATOR */
+
 extern int ext3_map_inode_page(struct inode *inode, struct page *page,
                                unsigned long *blocks, int *created, int create);
-int fsfilt_ext3_map_inode_page(struct inode *inode, struct page *page,
-                               unsigned long *blocks, int *created, int create)
+int fsfilt_ext3_map_bm_inode_pages(struct inode *inode, struct page **page,
+                                   int pages, unsigned long *blocks,
+                                   int *created, int create)
+{
+        int blocks_per_page = CFS_PAGE_SIZE >> inode->i_blkbits;
+        unsigned long *b;
+        int rc = 0, i, *cr;
+
+        for (i = 0, cr = created, b = blocks; i < pages; i++, page++) {
+                rc = ext3_map_inode_page(inode, *page, b, cr, create);
+                if (rc) {
+                        CERROR("ino %lu, blk %lu cr %u create %d: rc %d\n",
+                               inode->i_ino, *b, *cr, create, rc);
+                        break;
+                }
+
+                b += blocks_per_page;
+                cr += blocks_per_page;
+        }
+        return rc;
+}
+
+int fsfilt_ext3_map_inode_pages(struct inode *inode, struct page **page,
+                                int pages, unsigned long *blocks,
+                                int *created, int create,
+                                struct semaphore *optional_sem)
 {
-        return ext3_map_inode_page(inode, page, blocks, created, create);
+        int rc;
+#ifdef EXT3_MULTIBLOCK_ALLOCATOR
+        if (EXT3_I(inode)->i_flags & EXT3_EXTENTS_FL) {
+                rc = fsfilt_ext3_map_ext_inode_pages(inode, page, pages,
+                                                     blocks, created, create);
+                return rc;
+        }
+#endif
+        if (optional_sem != NULL)
+                down(optional_sem);
+        rc = fsfilt_ext3_map_bm_inode_pages(inode, page, pages, blocks,
+                                            created, create);
+        if (optional_sem != NULL)
+                up(optional_sem);
+
+        return rc;
 }
 
 extern int ext3_prep_san_write(struct inode *inode, long *blocks,
@@ -752,7 +1182,7 @@ static int fsfilt_ext3_write_record(struct file *file, void *buf, int bufsize,
         loff_t new_size = inode->i_size;
         journal_t *journal;
         handle_t *handle;
-        int err, block_count = 0, blocksize, size, boffs;
+        int err = 0, block_count = 0, blocksize, size, boffs;
 
         /* Determine how many transaction credits are needed */
         blocksize = 1 << inode->i_blkbits;
@@ -760,12 +1190,13 @@ static int fsfilt_ext3_write_record(struct file *file, void *buf, int bufsize,
         block_count = (block_count + blocksize - 1) >> inode->i_blkbits;
 
         journal = EXT3_SB(inode->i_sb)->s_journal;
-        lock_kernel();
+        lock_24kernel();
         handle = journal_start(journal,
-                               block_count * EXT3_DATA_TRANS_BLOCKS + 2);
-        unlock_kernel();
+                               block_count * FSFILT_DATA_TRANS_BLOCKS(inode->i_sb) + 2);
+        unlock_24kernel();
         if (IS_ERR(handle)) {
-                CERROR("can't start transaction\n");
+                CERROR("can't start transaction for %d blocks (%d bytes)\n",
+                       block_count * FSFILT_DATA_TRANS_BLOCKS(inode->i_sb) + 2, bufsize);
                 return PTR_ERR(handle);
         }
 
@@ -821,9 +1252,9 @@ out:
                 unlock_kernel();
         }
 
-        lock_kernel();
+        lock_24kernel();
         journal_stop(handle);
-        unlock_kernel();
+        unlock_24kernel();
 
         if (err == 0)
                 *offs = offset;
@@ -841,37 +1272,773 @@ static int fsfilt_ext3_setup(struct super_block *sb)
         set_opt(EXT3_SB(sb)->s_mount_opt, PDIROPS);
         sb->s_flags |= S_PDIROPS;
 #endif
+        if (!EXT3_HAS_COMPAT_FEATURE(sb, EXT3_FEATURE_COMPAT_DIR_INDEX))
+                CWARN("filesystem doesn't have dir_index feature enabled\n");
         return 0;
 }
 
+/* If fso is NULL, op is FSFILT operation, otherwise op is number of fso
+   objects. Logs is number of logfiles to update */
+static int fsfilt_ext3_get_op_len(int op, struct fsfilt_objinfo *fso, int logs)
+{
+        if ( !fso ) {
+                switch(op) {
+                case FSFILT_OP_CREATE:
+                                 /* directory leaf, index & indirect & EA*/
+                        return 4 + 3 * logs;
+                case FSFILT_OP_UNLINK:
+                        return 3 * logs;
+                }
+        } else {
+                int i;
+                int needed = 0;
+                struct super_block *sb = fso->fso_dentry->d_inode->i_sb;
+                int blockpp = 1 << (CFS_PAGE_SHIFT - sb->s_blocksize_bits);
+                int addrpp = EXT3_ADDR_PER_BLOCK(sb) * blockpp;
+                for (i = 0; i < op; i++, fso++) {
+                        int nblocks = fso->fso_bufcnt * blockpp;
+                        int ndindirect = min(nblocks, addrpp + 1);
+                        int nindir = nblocks + ndindirect + 1;
+
+                        needed += nindir;
+                }
+                return needed + 3 * logs;
+        }
+
+        return 0;
+}
+
+static const char *op_quotafile[] = { "lquota.user", "lquota.group" };
+
+#define DQINFO_COPY(out, in)                    \
+do {                                            \
+        Q_COPY(out, in, dqi_bgrace);            \
+        Q_COPY(out, in, dqi_igrace);            \
+        Q_COPY(out, in, dqi_flags);             \
+        Q_COPY(out, in, dqi_valid);             \
+} while (0)
+
+#define DQBLK_COPY(out, in)                     \
+do {                                            \
+        Q_COPY(out, in, dqb_bhardlimit);        \
+        Q_COPY(out, in, dqb_bsoftlimit);        \
+        Q_COPY(out, in, dqb_curspace);          \
+        Q_COPY(out, in, dqb_ihardlimit);        \
+        Q_COPY(out, in, dqb_isoftlimit);        \
+        Q_COPY(out, in, dqb_curinodes);         \
+        Q_COPY(out, in, dqb_btime);             \
+        Q_COPY(out, in, dqb_itime);             \
+        Q_COPY(out, in, dqb_valid);             \
+} while (0)
+
+      
+
+static int fsfilt_ext3_quotactl(struct super_block *sb,
+                                struct obd_quotactl *oqc)
+{
+        int i, rc = 0, error = 0;
+        struct quotactl_ops *qcop;
+        struct if_dqinfo *info;
+        struct if_dqblk *dqblk;
+        ENTRY;
+
+        if (!sb->s_qcop)
+                RETURN(-ENOSYS);
+
+        OBD_ALLOC_PTR(info);
+        if (!info)
+                RETURN(-ENOMEM);
+        OBD_ALLOC_PTR(dqblk);
+        if (!dqblk) {
+                OBD_FREE_PTR(info);
+                RETURN(-ENOMEM);
+        }
+
+        DQINFO_COPY(info, &oqc->qc_dqinfo);
+        DQBLK_COPY(dqblk, &oqc->qc_dqblk);
+
+        qcop = sb->s_qcop;
+        if (oqc->qc_cmd == Q_QUOTAON || oqc->qc_cmd == Q_QUOTAOFF) {
+                for (i = 0; i < MAXQUOTAS; i++) {
+                        if (!Q_TYPESET(oqc, i))
+                                continue;
+
+                        if (oqc->qc_cmd == Q_QUOTAON) {
+                                if (!qcop->quota_on)
+                                        GOTO(out, rc = -ENOSYS);
+                                rc = qcop->quota_on(sb, i, oqc->qc_id,
+                                                    (char *)op_quotafile[i]);
+                        } else if (oqc->qc_cmd == Q_QUOTAOFF) {
+                                if (!qcop->quota_off)
+                                        GOTO(out, rc = -ENOSYS);
+                                rc = qcop->quota_off(sb, i);
+                        }
+
+                        if (rc == -EBUSY)
+                                error = rc;
+                        else if (rc)
+                                GOTO(out, rc);
+                }
+                GOTO(out, rc ?: error);
+        }
+
+        switch (oqc->qc_cmd) {
+        case Q_GETOINFO:
+        case Q_GETINFO:
+                if (!qcop->get_info)
+                        GOTO(out, rc = -ENOSYS);
+                rc = qcop->get_info(sb, oqc->qc_type, info);
+                break;
+        case Q_SETQUOTA:
+        case Q_INITQUOTA:
+                if (!qcop->set_dqblk)
+                        GOTO(out, rc = -ENOSYS);
+                rc = qcop->set_dqblk(sb, oqc->qc_type, oqc->qc_id, dqblk);
+                break;
+        case Q_GETOQUOTA:
+        case Q_GETQUOTA:
+                if (!qcop->get_dqblk)
+                        GOTO(out, rc = -ENOSYS);
+                rc = qcop->get_dqblk(sb, oqc->qc_type, oqc->qc_id, dqblk);
+                break;
+        case Q_SYNC:
+                if (!sb->s_qcop->quota_sync)
+                        GOTO(out, rc = -ENOSYS);
+                qcop->quota_sync(sb, oqc->qc_type);
+                break;
+        default:
+                CERROR("unsupported quotactl command: %d", oqc->qc_cmd);
+                LBUG();
+        }
+out:
+        DQINFO_COPY(&oqc->qc_dqinfo, info);
+        DQBLK_COPY(&oqc->qc_dqblk, dqblk);
+
+        OBD_FREE_PTR(info);
+        OBD_FREE_PTR(dqblk);
+
+        if (rc)
+                CDEBUG(D_QUOTA, "quotactl command %#x, id %u, type %d "
+                                "failed: %d\n",
+                       oqc->qc_cmd, oqc->qc_id, oqc->qc_type, rc);
+        RETURN(rc);
+}
+
+struct chk_dqblk{
+        struct hlist_node       dqb_hash;        /* quotacheck hash */
+        struct list_head        dqb_list;        /* in list also */
+        qid_t                   dqb_id;          /* uid/gid */
+        short                   dqb_type;        /* USRQUOTA/GRPQUOTA */
+        __u32                   dqb_bhardlimit;  /* block hard limit */
+        __u32                   dqb_bsoftlimit;  /* block soft limit */
+        qsize_t                 dqb_curspace;    /* current space */
+        __u32                   dqb_ihardlimit;  /* inode hard limit */
+        __u32                   dqb_isoftlimit;  /* inode soft limit */
+        __u32                   dqb_curinodes;   /* current inodes */
+        __u64                   dqb_btime;       /* block grace time */
+        __u64                   dqb_itime;       /* inode grace time */
+        __u32                   dqb_valid;       /* flag for above fields */
+};
+
+static inline unsigned int chkquot_hash(qid_t id, int type)
+                                        __attribute__((__const__));
+
+static inline unsigned int chkquot_hash(qid_t id, int type)
+{
+        return (id * (MAXQUOTAS - type)) % NR_DQHASH;
+}
+
+static inline struct chk_dqblk *
+find_chkquot(struct hlist_head *head, qid_t id, int type)
+{
+        struct hlist_node *node;
+        struct chk_dqblk *cdqb;
+
+        hlist_for_each(node, head) {
+                cdqb = hlist_entry(node, struct chk_dqblk, dqb_hash);
+                if (cdqb->dqb_id == id && cdqb->dqb_type == type)
+                        return cdqb;
+        }
+
+        return NULL;
+}
+
+static struct chk_dqblk *alloc_chkquot(qid_t id, int type)
+{
+        struct chk_dqblk *cdqb;
+
+        OBD_ALLOC_PTR(cdqb);
+        if (cdqb) {
+                INIT_HLIST_NODE(&cdqb->dqb_hash);
+                INIT_LIST_HEAD(&cdqb->dqb_list);
+                cdqb->dqb_id = id;
+                cdqb->dqb_type = type;
+        }
+
+        return cdqb;
+}
+
+static struct chk_dqblk *
+cqget(struct super_block *sb, struct hlist_head *hash, struct list_head *list,
+      qid_t id, int type, int first_check)
+{
+        struct hlist_head *head = hash + chkquot_hash(id, type);
+        struct if_dqblk dqb;
+        struct chk_dqblk *cdqb;
+        int rc;
+
+        cdqb = find_chkquot(head, id, type);
+        if (cdqb)
+                return cdqb;
+
+        cdqb = alloc_chkquot(id, type);
+        if (!cdqb)
+                return NULL;
+
+        if (!first_check) {
+                rc = sb->s_qcop->get_dqblk(sb, type, id, &dqb);
+                if (rc) {
+                        CERROR("get_dqblk of id %u, type %d failed: %d\n",
+                               id, type, rc);
+                } else {
+                        DQBLK_COPY(cdqb, &dqb);
+                        cdqb->dqb_curspace = 0;
+                        cdqb->dqb_curinodes = 0;
+                }
+        }
+
+        hlist_add_head(&cdqb->dqb_hash, head);
+        list_add_tail(&cdqb->dqb_list, list);
+
+        return cdqb;
+}
+
+static inline int quota_onoff(struct super_block *sb, int cmd, int type)
+{
+        struct obd_quotactl *oqctl;
+        int rc;
+
+        OBD_ALLOC_PTR(oqctl);
+        if (!oqctl)
+                RETURN(-ENOMEM);
+
+        oqctl->qc_cmd = cmd;
+        oqctl->qc_id = QFMT_LDISKFS;
+        oqctl->qc_type = type;
+        rc = fsfilt_ext3_quotactl(sb, oqctl);
+
+        OBD_FREE_PTR(oqctl);
+        return rc;
+}
+
+static inline int read_old_dqinfo(struct super_block *sb, int type,
+                                  struct if_dqinfo *dqinfo)
+{
+        struct obd_quotactl *oqctl;
+        int rc;
+        ENTRY;
+
+        OBD_ALLOC_PTR(oqctl);
+        if (!oqctl)
+                RETURN(-ENOMEM);
+
+        oqctl->qc_cmd = Q_GETINFO;
+        oqctl->qc_type = type;
+        rc = fsfilt_ext3_quotactl(sb, oqctl);
+        if (!rc)
+                ((struct obd_dqinfo *)dqinfo)[type] = oqctl->qc_dqinfo;
+
+        OBD_FREE_PTR(oqctl);
+        RETURN(rc);
+}
+
+static inline struct ext3_group_desc *
+get_group_desc(struct super_block *sb, int group)
+{
+        unsigned long desc_block, desc;
+        struct ext3_group_desc *gdp;
+
+        desc_block = group / EXT3_DESC_PER_BLOCK(sb);
+        desc = group % EXT3_DESC_PER_BLOCK(sb);
+        gdp = (struct ext3_group_desc *)
+              EXT3_SB(sb)->s_group_desc[desc_block]->b_data;
+
+        return gdp + desc;
+}
+
+static inline struct buffer_head *
+read_inode_bitmap(struct super_block *sb, unsigned long group)
+{
+        struct ext3_group_desc *desc;
+        struct buffer_head *bh;
+
+        desc = get_group_desc(sb, group);
+        bh = sb_bread(sb, le32_to_cpu(desc->bg_inode_bitmap));
+
+        return bh;
+}
+
+static inline struct inode *ext3_iget_inuse(struct super_block *sb,
+                                     struct buffer_head *bitmap_bh,
+                                     int index, unsigned long ino)
+{
+        struct inode *inode = NULL;
+
+        if (ext3_test_bit(index, bitmap_bh->b_data))
+                inode = iget(sb, ino);
+
+        return inode;
+}
+
+struct qchk_ctxt {
+        struct hlist_head       qckt_hash[NR_DQHASH];        /* quotacheck hash */
+        struct list_head        qckt_list;                   /* quotacheck list */
+        int                     qckt_first_check[MAXQUOTAS]; /* 1 if no old quotafile */
+        struct if_dqinfo        qckt_dqinfo[MAXQUOTAS];      /* old dqinfo */
+};
+
+static int add_inode_quota(struct inode *inode, struct qchk_ctxt *qctxt,
+                           struct obd_quotactl *oqc)
+{
+        struct chk_dqblk *cdqb[MAXQUOTAS] = { NULL, };
+        loff_t size = 0;
+        qid_t qid[MAXQUOTAS];
+        int cnt, i, rc = 0;
+
+        if (!inode)
+                return 0;
+
+        qid[USRQUOTA] = inode->i_uid;
+        qid[GRPQUOTA] = inode->i_gid;
+
+        if (S_ISDIR(inode->i_mode) ||
+            S_ISREG(inode->i_mode) ||
+            S_ISLNK(inode->i_mode))
+                size = inode_get_bytes(inode);
+
+        for (cnt = 0; cnt < MAXQUOTAS; cnt++) {
+                if (!Q_TYPESET(oqc, cnt))
+                        continue;
+
+                cdqb[cnt] = cqget(inode->i_sb, qctxt->qckt_hash,
+                                &qctxt->qckt_list, qid[cnt], cnt,
+                                qctxt->qckt_first_check[cnt]);
+                if (!cdqb[cnt]) {
+                        rc = -ENOMEM;
+                        break;
+                }
+
+                cdqb[cnt]->dqb_curspace += size;
+                cdqb[cnt]->dqb_curinodes++;
+        }
+
+        if (rc) {
+                for (i = 0; i < cnt; i++) {
+                        if (!Q_TYPESET(oqc, i))
+                                continue;
+                        LASSERT(cdqb[i]);
+                        cdqb[i]->dqb_curspace -= size;
+                        cdqb[i]->dqb_curinodes--;
+                }
+        }
+
+        return rc;
+}
+
+static int v2_write_dqheader(struct file *f, int type)
+{
+        static const __u32 quota_magics[] = V2_INITQMAGICS;
+        static const __u32 quota_versions[] = V2_INITQVERSIONS;
+        struct v2_disk_dqheader dqhead;
+        loff_t offset = 0;
+
+        CLASSERT(ARRAY_SIZE(quota_magics) == ARRAY_SIZE(quota_versions));
+        LASSERT(0 <= type && type < ARRAY_SIZE(quota_magics));
+
+        dqhead.dqh_magic = cpu_to_le32(quota_magics[type]);
+        dqhead.dqh_version = cpu_to_le32(quota_versions[type]);
+
+        return cfs_user_write(f, (char *)&dqhead, sizeof(dqhead), &offset);
+}
+
+/* write dqinfo struct in a new quota file */
+static int v2_write_dqinfo(struct file *f, int type, struct if_dqinfo *info)
+{
+        struct v2_disk_dqinfo dqinfo;
+        __u32 blocks = V2_DQTREEOFF + 1;
+        loff_t offset = V2_DQINFOOFF;
+
+        if (info) {
+                dqinfo.dqi_bgrace = cpu_to_le32(info->dqi_bgrace);
+                dqinfo.dqi_igrace = cpu_to_le32(info->dqi_igrace);
+                dqinfo.dqi_flags = cpu_to_le32(info->dqi_flags & DQF_MASK &
+                                               ~DQF_INFO_DIRTY);
+        } else {
+                dqinfo.dqi_bgrace = cpu_to_le32(MAX_DQ_TIME);
+                dqinfo.dqi_igrace = cpu_to_le32(MAX_IQ_TIME);
+                dqinfo.dqi_flags = 0;
+        }
+
+        dqinfo.dqi_blocks = cpu_to_le32(blocks);
+        dqinfo.dqi_free_blk = 0;
+        dqinfo.dqi_free_entry = 0;
+
+        return cfs_user_write(f, (char *)&dqinfo, sizeof(dqinfo), &offset);
+}
+
+static int create_new_quota_files(struct qchk_ctxt *qctxt,
+                                  struct obd_quotactl *oqc)
+{
+        int i, rc = 0;
+        ENTRY;
+
+        for (i = 0; i < MAXQUOTAS; i++) {
+                struct if_dqinfo *info = qctxt->qckt_first_check[i]?
+                                         NULL : &qctxt->qckt_dqinfo[i];
+                struct file *file;
+
+                if (!Q_TYPESET(oqc, i))
+                        continue;
+
+                file = filp_open(op_quotafile[i], O_RDWR | O_CREAT | O_TRUNC,
+                                 0644);
+                if (IS_ERR(file)) {
+                        rc = PTR_ERR(file);
+                        CERROR("can't create %s file: rc = %d\n",
+                               op_quotafile[i], rc);
+                        GOTO(out, rc);
+                }
+
+                if (!S_ISREG(file->f_dentry->d_inode->i_mode)) {
+                        CERROR("file %s is not regular", op_quotafile[i]);
+                        filp_close(file, 0);
+                        GOTO(out, rc = -EINVAL);
+                }
+
+                rc = v2_write_dqheader(file, i);
+                if (rc) {
+                        filp_close(file, 0);
+                        GOTO(out, rc);
+                }
+
+                rc = v2_write_dqinfo(file, i, info);
+                filp_close(file, 0);
+                if (rc)
+                        GOTO(out, rc);
+        }
+
+out:
+        RETURN(rc);
+}
+
+
+static int commit_chkquot(struct super_block *sb, struct qchk_ctxt *qctxt,
+                          struct chk_dqblk *cdqb)
+{
+        struct obd_quotactl *oqc;
+        long now;
+        int rc;
+        ENTRY;
+
+        OBD_ALLOC_PTR(oqc);
+        if (!oqc)
+                RETURN(-ENOMEM);
+
+        now = CURRENT_SECONDS;
+
+        if (cdqb->dqb_bsoftlimit &&
+            toqb(cdqb->dqb_curspace) >= cdqb->dqb_bsoftlimit &&
+            !cdqb->dqb_btime)
+                cdqb->dqb_btime =
+                        now + qctxt->qckt_dqinfo[cdqb->dqb_type].dqi_bgrace;
+
+        if (cdqb->dqb_isoftlimit &&
+            cdqb->dqb_curinodes >= cdqb->dqb_isoftlimit &&
+            !cdqb->dqb_itime)
+                cdqb->dqb_itime =
+                        now + qctxt->qckt_dqinfo[cdqb->dqb_type].dqi_igrace;
+
+        cdqb->dqb_valid = QIF_ALL;
+
+        oqc->qc_cmd = Q_SETQUOTA;
+        oqc->qc_type = cdqb->dqb_type;
+        oqc->qc_id = cdqb->dqb_id;
+        DQBLK_COPY(&oqc->qc_dqblk, cdqb);
+
+        rc = fsfilt_ext3_quotactl(sb, oqc);
+        OBD_FREE_PTR(oqc);
+        RETURN(rc);
+}
+
+static int prune_chkquots(struct super_block *sb,
+                          struct qchk_ctxt *qctxt, int error)
+{
+        struct chk_dqblk *cdqb, *tmp;
+        int rc;
+
+        list_for_each_entry_safe(cdqb, tmp, &qctxt->qckt_list, dqb_list) {
+                if (!error) {
+                        rc = commit_chkquot(sb, qctxt, cdqb);
+                        if (rc)
+                                error = rc;
+                }
+                hlist_del_init(&cdqb->dqb_hash);
+                list_del(&cdqb->dqb_list);
+                OBD_FREE_PTR(cdqb);
+        }
+
+        return error;
+}
+
+static int fsfilt_ext3_quotacheck(struct super_block *sb,
+                                  struct obd_quotactl *oqc)
+{
+        struct ext3_sb_info *sbi = EXT3_SB(sb);
+        int i, group;
+        struct qchk_ctxt *qctxt;
+        struct buffer_head *bitmap_bh = NULL;
+        unsigned long ino;
+        struct inode *inode;
+        int rc = 0;
+        ENTRY;
+
+        /* turn on quota and read dqinfo if existed */
+        OBD_ALLOC_PTR(qctxt);
+        if (!qctxt) {
+                oqc->qc_stat = -ENOMEM;
+                RETURN(-ENOMEM);
+        }
+
+        for (i = 0; i < NR_DQHASH; i++)
+                INIT_HLIST_HEAD(&qctxt->qckt_hash[i]);
+        INIT_LIST_HEAD(&qctxt->qckt_list);
+
+        for (i = 0; i < MAXQUOTAS; i++) {
+                if (!Q_TYPESET(oqc, i))
+                        continue;
+
+                rc = quota_onoff(sb, Q_QUOTAON, i);
+                if (!rc || rc == -EBUSY) {
+                        rc = read_old_dqinfo(sb, i, qctxt->qckt_dqinfo);
+                        if (rc)
+                                GOTO(out, rc);
+                } else if (rc == -ENOENT) {
+                        qctxt->qckt_first_check[i] = 1;
+                } else if (rc) {
+                        GOTO(out, rc);
+                }
+        }
+
+        /* check quota and update in hash */
+        for (group = 0; group < sbi->s_groups_count; group++) {
+                ino = group * sbi->s_inodes_per_group + 1;
+                bitmap_bh = read_inode_bitmap(sb, group);
+                if (!bitmap_bh) {
+                        CERROR("read_inode_bitmap group %d failed", group);
+                        GOTO(out, -EIO);
+                }
+
+                for (i = 0; i < sbi->s_inodes_per_group; i++, ino++) {
+                        if (ino < sbi->s_first_ino)
+                                continue;
+
+                        inode = ext3_iget_inuse(sb, bitmap_bh, i, ino);
+                        rc = add_inode_quota(inode, qctxt, oqc);
+                        iput(inode);
+                        if (rc) {
+                                brelse(bitmap_bh);
+                                GOTO(out, rc);
+                        }
+                }
+
+                brelse(bitmap_bh);
+        }
+
+        /* read old quota limits from old quota file. (only for the user
+         * has limits but hasn't file) */
+#ifdef HAVE_QUOTA_SUPPORT
+        for (i = 0; i < MAXQUOTAS; i++) {
+                struct list_head id_list;
+                struct dquot_id *dqid, *tmp;
+
+                if (!Q_TYPESET(oqc, i))
+                        continue;
+
+                if (qctxt->qckt_first_check[i])
+                        continue;
+
+
+                LASSERT(sb_dqopt(sb)->files[i] != NULL);
+                INIT_LIST_HEAD(&id_list);
+#ifndef KERNEL_SUPPORTS_QUOTA_READ 
+                rc = lustre_get_qids(sb_dqopt(sb)->files[i], NULL, i, &id_list);
+#else
+                rc = lustre_get_qids(NULL, sb_dqopt(sb)->files[i], i, &id_list);
+#endif
+                if (rc)
+                        CERROR("read old limits failed. (rc:%d)\n", rc);
+
+                list_for_each_entry_safe(dqid, tmp, &id_list, di_link) {
+                        list_del_init(&dqid->di_link);
+
+                        if (!rc)
+                                cqget(sb, qctxt->qckt_hash, &qctxt->qckt_list,
+                                      dqid->di_id, i,
+                                      qctxt->qckt_first_check[i]);
+                        kfree(dqid);
+                }
+        }
+#endif
+        /* turn off quota cause we are to dump chk_dqblk to files */
+        quota_onoff(sb, Q_QUOTAOFF, oqc->qc_type);
+
+        rc = create_new_quota_files(qctxt, oqc);
+        if (rc)
+                GOTO(out, rc);
+
+        /* we use vfs functions to set dqblk, so turn quota on */
+        rc = quota_onoff(sb, Q_QUOTAON, oqc->qc_type);
+out:
+        /* dump and free chk_dqblk */
+        rc = prune_chkquots(sb, qctxt, rc);
+        OBD_FREE_PTR(qctxt);
+
+        /* turn off quota, `lfs quotacheck` will turn on when all
+         * nodes quotacheck finish. */
+        quota_onoff(sb, Q_QUOTAOFF, oqc->qc_type);
+
+        oqc->qc_stat = rc;
+        if (rc)
+                CERROR("quotacheck failed: rc = %d\n", rc);
+
+        RETURN(rc);
+}
+
+#ifdef HAVE_QUOTA_SUPPORT
+static int fsfilt_ext3_quotainfo(struct lustre_quota_info *lqi, int type, 
+                                 int cmd)
+{
+        int rc = 0;
+        ENTRY;
+
+        if (lqi->qi_files[type] == NULL) {
+                CERROR("operate qinfo before it's enabled!\n");
+                RETURN(-EIO);
+        }
+
+        switch (cmd) {
+        case QFILE_CHK:
+                rc = lustre_check_quota_file(lqi, type);
+                break;
+        case QFILE_RD_INFO:
+                rc = lustre_read_quota_info(lqi, type);
+                break;
+        case QFILE_WR_INFO:
+                rc = lustre_write_quota_info(lqi, type);
+                break;
+        case QFILE_INIT_INFO:
+                rc = lustre_init_quota_info(lqi, type);
+                break;
+        default:
+                CERROR("Unsupported admin quota file cmd %d\n", cmd);
+                LBUG();
+                break;
+        }
+        RETURN(rc);
+}
+
+static int fsfilt_ext3_qids(struct file *file, struct inode *inode, int type,
+                            struct list_head *list)
+{
+        return lustre_get_qids(file, inode, type, list);
+}
+
+static int fsfilt_ext3_dquot(struct lustre_dquot *dquot, int cmd)
+{
+        int rc = 0;
+        ENTRY;
+
+        if (dquot->dq_info->qi_files[dquot->dq_type] == NULL) {
+                CERROR("operate dquot before it's enabled!\n");
+                RETURN(-EIO);
+        }
+
+        switch (cmd) {
+        case QFILE_RD_DQUOT:
+                rc = lustre_read_dquot(dquot);
+                break;
+        case QFILE_WR_DQUOT:
+                if (dquot->dq_dqb.dqb_ihardlimit ||
+                    dquot->dq_dqb.dqb_isoftlimit ||
+                    dquot->dq_dqb.dqb_bhardlimit ||
+                    dquot->dq_dqb.dqb_bsoftlimit)
+                        clear_bit(DQ_FAKE_B, &dquot->dq_flags);
+                else
+                        set_bit(DQ_FAKE_B, &dquot->dq_flags);
+
+                rc = lustre_commit_dquot(dquot);
+                if (rc >= 0)
+                        rc = 0;
+                break;
+        default:
+                CERROR("Unsupported admin quota file cmd %d\n", cmd);
+                LBUG();
+                break;
+        }
+        RETURN(rc);
+}
+#endif
+
+static lvfs_sbdev_type fsfilt_ext3_journal_sbdev(struct super_block *sb)
+{
+#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2,5,0))
+        return (EXT3_SB(sb)->journal_bdev);
+#else
+        return kdev_t_to_nr(EXT3_SB(sb)->s_journal->j_dev);
+#endif
+}
+
 static struct fsfilt_operations fsfilt_ext3_ops = {
-        fs_type:                "ext3",
-        fs_owner:               THIS_MODULE,
-        fs_start:               fsfilt_ext3_start,
-        fs_brw_start:           fsfilt_ext3_brw_start,
-        fs_commit:              fsfilt_ext3_commit,
-        fs_commit_async:        fsfilt_ext3_commit_async,
-        fs_commit_wait:         fsfilt_ext3_commit_wait,
-        fs_setattr:             fsfilt_ext3_setattr,
-        fs_iocontrol:           fsfilt_ext3_iocontrol,
-        fs_set_md:              fsfilt_ext3_set_md,
-        fs_get_md:              fsfilt_ext3_get_md,
-        fs_readpage:            fsfilt_ext3_readpage,
-        fs_add_journal_cb:      fsfilt_ext3_add_journal_cb,
-        fs_statfs:              fsfilt_ext3_statfs,
-        fs_sync:                fsfilt_ext3_sync,
-        fs_map_inode_page:      fsfilt_ext3_map_inode_page,
-        fs_prep_san_write:      fsfilt_ext3_prep_san_write,
-        fs_write_record:        fsfilt_ext3_write_record,
-        fs_read_record:         fsfilt_ext3_read_record,
-        fs_setup:               fsfilt_ext3_setup,
+        .fs_type                = "ext3",
+        .fs_owner               = THIS_MODULE,
+        .fs_getlabel            = fsfilt_ext3_get_label,
+        .fs_setlabel            = fsfilt_ext3_set_label,
+        .fs_uuid                = fsfilt_ext3_uuid,
+        .fs_start               = fsfilt_ext3_start,
+        .fs_brw_start           = fsfilt_ext3_brw_start,
+        .fs_extend              = fsfilt_ext3_extend,
+        .fs_commit              = fsfilt_ext3_commit,
+        .fs_commit_async        = fsfilt_ext3_commit_async,
+        .fs_commit_wait         = fsfilt_ext3_commit_wait,
+        .fs_setattr             = fsfilt_ext3_setattr,
+        .fs_iocontrol           = fsfilt_ext3_iocontrol,
+        .fs_set_md              = fsfilt_ext3_set_md,
+        .fs_get_md              = fsfilt_ext3_get_md,
+        .fs_readpage            = fsfilt_ext3_readpage,
+        .fs_add_journal_cb      = fsfilt_ext3_add_journal_cb,
+        .fs_statfs              = fsfilt_ext3_statfs,
+        .fs_sync                = fsfilt_ext3_sync,
+        .fs_map_inode_pages     = fsfilt_ext3_map_inode_pages,
+        .fs_prep_san_write      = fsfilt_ext3_prep_san_write,
+        .fs_write_record        = fsfilt_ext3_write_record,
+        .fs_read_record         = fsfilt_ext3_read_record,
+        .fs_setup               = fsfilt_ext3_setup,
+        .fs_send_bio            = fsfilt_ext3_send_bio,
+        .fs_get_op_len          = fsfilt_ext3_get_op_len,
+        .fs_quotactl            = fsfilt_ext3_quotactl,
+        .fs_quotacheck          = fsfilt_ext3_quotacheck,
+#ifdef HAVE_QUOTA_SUPPORT
+        .fs_quotainfo           = fsfilt_ext3_quotainfo,
+        .fs_qids                = fsfilt_ext3_qids,
+        .fs_dquot               = fsfilt_ext3_dquot,
+#endif
+        .fs_journal_sbdev       = fsfilt_ext3_journal_sbdev,
 };
 
 static int __init fsfilt_ext3_init(void)
 {
         int rc;
 
-        //rc = ext3_xattr_register();
         fcb_cache = kmem_cache_create("fsfilt_ext3_fcb",
                                       sizeof(struct fsfilt_cb_data), 0,
                                       0, NULL, NULL);
@@ -882,8 +2049,10 @@ static int __init fsfilt_ext3_init(void)
 
         rc = fsfilt_register_ops(&fsfilt_ext3_ops);
 
-        if (rc)
-                kmem_cache_destroy(fcb_cache);
+        if (rc) {
+                int err = kmem_cache_destroy(fcb_cache);
+                LASSERTF(err == 0, "error destroying new cache: rc %d\n", err);
+        }
 out:
         return rc;
 }
@@ -894,13 +2063,7 @@ static void __exit fsfilt_ext3_exit(void)
 
         fsfilt_unregister_ops(&fsfilt_ext3_ops);
         rc = kmem_cache_destroy(fcb_cache);
-
-        if (rc || atomic_read(&fcb_cache_count)) {
-                CERROR("can't free fsfilt callback cache: count %d, rc = %d\n",
-                       atomic_read(&fcb_cache_count), rc);
-        }
-
-        //rc = ext3_xattr_unregister();
+        LASSERTF(rc == 0, "couldn't destroy fcb_cache slab\n");
 }
 
 module_init(fsfilt_ext3_init);