-Index: linux-2.6.9-full/include/linux/ext3_fs_sb.h
-===================================================================
---- linux-2.6.9-full.orig/include/linux/ext3_fs_sb.h 2006-05-18 23:57:04.000000000 +0400
-+++ linux-2.6.9-full/include/linux/ext3_fs_sb.h 2007-03-28 15:42:16.000000000 +0400
-@@ -81,6 +81,61 @@ struct ext3_sb_info {
- char *s_qf_names[MAXQUOTAS]; /* Names of quota files with journalled quota */
- int s_jquota_fmt; /* Format of quota to use */
- #endif
-+
-+ /* for buddy allocator */
-+ struct ext3_group_info ***s_group_info;
-+ struct inode *s_buddy_cache;
-+ long s_blocks_reserved;
-+ spinlock_t s_reserve_lock;
-+ struct list_head s_active_transaction;
-+ struct list_head s_closed_transaction;
-+ struct list_head s_committed_transaction;
-+ spinlock_t s_md_lock;
-+ tid_t s_last_transaction;
-+ unsigned short *s_mb_offsets, *s_mb_maxs;
-+
-+ /* tunables */
-+ unsigned long s_mb_factor;
-+ unsigned long s_stripe;
-+ unsigned long s_mb_stream_request;
-+ unsigned long s_mb_max_to_scan;
-+ unsigned long s_mb_min_to_scan;
-+ unsigned long s_mb_max_groups_to_scan;
-+ unsigned long s_mb_stats;
-+ unsigned long s_mb_order2_reqs;
-+
-+ /* history to debug policy */
-+ struct ext3_mb_history *s_mb_history;
-+ int s_mb_history_cur;
-+ int s_mb_history_max;
-+ int s_mb_history_num;
-+ struct proc_dir_entry *s_mb_proc;
-+ spinlock_t s_mb_history_lock;
-+ int s_mb_history_filter;
-+
-+ /* stats for buddy allocator */
-+ spinlock_t s_mb_pa_lock;
-+ atomic_t s_bal_reqs; /* number of reqs with len > 1 */
-+ atomic_t s_bal_success; /* we found long enough chunks */
-+ atomic_t s_bal_allocated; /* in blocks */
-+ atomic_t s_bal_ex_scanned; /* total extents scanned */
-+ atomic_t s_bal_goals; /* goal hits */
-+ atomic_t s_bal_breaks; /* too long searches */
-+ atomic_t s_bal_2orders; /* 2^order hits */
-+ spinlock_t s_bal_lock;
-+ unsigned long s_mb_buddies_generated;
-+ unsigned long long s_mb_generation_time;
-+ atomic_t s_mb_lost_chunks;
-+ atomic_t s_mb_preallocated;
-+ atomic_t s_mb_discarded;
-+
-+ /* locality groups */
-+ struct ext3_locality_group *s_locality_groups;
-+
- };
-
-+#define EXT3_GROUP_INFO(sb, group) \
-+ EXT3_SB(sb)->s_group_info[(group) >> EXT3_DESC_PER_BLOCK_BITS(sb)] \
-+ [(group) & (EXT3_DESC_PER_BLOCK(sb) - 1)]
-+
- #endif /* _LINUX_EXT3_FS_SB */
Index: linux-2.6.9-full/include/linux/ext3_fs.h
===================================================================
---- linux-2.6.9-full.orig/include/linux/ext3_fs.h 2007-03-28 01:29:39.000000000 +0400
-+++ linux-2.6.9-full/include/linux/ext3_fs.h 2007-03-28 15:45:07.000000000 +0400
+--- linux-2.6.9-full.orig/include/linux/ext3_fs.h 2007-06-08 23:44:08.000000000 +0400
++++ linux-2.6.9-full/include/linux/ext3_fs.h 2007-10-17 22:25:01.000000000 +0400
@@ -57,6 +57,30 @@ struct statfs;
#define ext3_debug(f, a...) do {} while (0)
#endif
/*
* Special inodes numbers
*/
-@@ -404,6 +413,14 @@
+@@ -387,6 +411,14 @@ struct ext3_inode {
#define ext3_find_first_zero_bit ext2_find_first_zero_bit
#define ext3_find_next_zero_bit ext2_find_next_zero_bit
/*
* Maximal mount counts between two filesystem checks
*/
-@@ -763,6 +787,20 @@ extern unsigned long ext3_count_dirs (st
+@@ -763,6 +795,20 @@ extern unsigned long ext3_count_dirs (st
extern void ext3_check_inodes_bitmap (struct super_block *);
extern unsigned long ext3_count_free (struct buffer_head *, unsigned);
/* inode.c */
extern int ext3_block_truncate_page(handle_t *, struct page *,
+Index: linux-2.6.9-full/include/linux/ext3_fs_sb.h
+===================================================================
+--- linux-2.6.9-full.orig/include/linux/ext3_fs_sb.h 2007-06-08 23:44:07.000000000 +0400
++++ linux-2.6.9-full/include/linux/ext3_fs_sb.h 2007-10-17 22:25:01.000000000 +0400
+@@ -81,6 +81,61 @@ struct ext3_sb_info {
+ char *s_qf_names[MAXQUOTAS]; /* Names of quota files with journalled quota */
+ int s_jquota_fmt; /* Format of quota to use */
+ #endif
++
++ /* for buddy allocator */
++ struct ext3_group_info ***s_group_info;
++ struct inode *s_buddy_cache;
++ long s_blocks_reserved;
++ spinlock_t s_reserve_lock;
++ struct list_head s_active_transaction;
++ struct list_head s_closed_transaction;
++ struct list_head s_committed_transaction;
++ spinlock_t s_md_lock;
++ tid_t s_last_transaction;
++ unsigned short *s_mb_offsets, *s_mb_maxs;
++
++ /* tunables */
++ unsigned long s_mb_factor;
++ unsigned long s_stripe;
++ unsigned long s_mb_stream_request;
++ unsigned long s_mb_max_to_scan;
++ unsigned long s_mb_min_to_scan;
++ unsigned long s_mb_max_groups_to_scan;
++ unsigned long s_mb_stats;
++ unsigned long s_mb_order2_reqs;
++
++ /* history to debug policy */
++ struct ext3_mb_history *s_mb_history;
++ int s_mb_history_cur;
++ int s_mb_history_max;
++ int s_mb_history_num;
++ struct proc_dir_entry *s_mb_proc;
++ spinlock_t s_mb_history_lock;
++ int s_mb_history_filter;
++
++ /* stats for buddy allocator */
++ spinlock_t s_mb_pa_lock;
++ atomic_t s_bal_reqs; /* number of reqs with len > 1 */
++ atomic_t s_bal_success; /* we found long enough chunks */
++ atomic_t s_bal_allocated; /* in blocks */
++ atomic_t s_bal_ex_scanned; /* total extents scanned */
++ atomic_t s_bal_goals; /* goal hits */
++ atomic_t s_bal_breaks; /* too long searches */
++ atomic_t s_bal_2orders; /* 2^order hits */
++ spinlock_t s_bal_lock;
++ unsigned long s_mb_buddies_generated;
++ unsigned long long s_mb_generation_time;
++ atomic_t s_mb_lost_chunks;
++ atomic_t s_mb_preallocated;
++ atomic_t s_mb_discarded;
++
++ /* locality groups */
++ struct ext3_locality_group *s_locality_groups;
++
+ };
+
++#define EXT3_GROUP_INFO(sb, group) \
++ EXT3_SB(sb)->s_group_info[(group) >> EXT3_DESC_PER_BLOCK_BITS(sb)] \
++ [(group) & (EXT3_DESC_PER_BLOCK(sb) - 1)]
++
+ #endif /* _LINUX_EXT3_FS_SB */
Index: linux-2.6.9-full/fs/ext3/super.c
===================================================================
---- linux-2.6.9-full.orig/fs/ext3/super.c 2007-03-28 01:29:38.000000000 +0400
-+++ linux-2.6.9-full/fs/ext3/super.c 2007-03-28 15:42:16.000000000 +0400
+--- linux-2.6.9-full.orig/fs/ext3/super.c 2007-06-08 23:44:08.000000000 +0400
++++ linux-2.6.9-full/fs/ext3/super.c 2007-10-17 22:26:27.000000000 +0400
@@ -394,6 +394,7 @@ void ext3_put_super (struct super_block
struct ext3_super_block *es = sbi->s_es;
int i;
journal_destroy(sbi->s_journal);
@@ -463,6 +464,8 @@ static struct inode *ext3_alloc_inode(st
ei->vfs_inode.i_version = 1;
-
+
memset(&ei->i_cached_extent, 0, sizeof(ei->i_cached_extent));
+ INIT_LIST_HEAD(&ei->i_prealloc_list);
+ spin_lock_init(&ei->i_prealloc_lock);
return &ei->vfs_inode;
}
-@@ -2433,7 +2436,13 @@ static struct file_system_type ext3_fs_t
+@@ -2576,7 +2579,13 @@ static struct file_system_type ext3_fs_t
static int __init init_ext3_fs(void)
{
if (err)
return err;
err = init_inodecache();
-@@ -2455,6 +2464,7 @@ static void __exit exit_ext3_fs(void)
+@@ -2598,6 +2607,7 @@ static void __exit exit_ext3_fs(void)
unregister_filesystem(&ext3_fs_type);
destroy_inodecache();
exit_ext3_xattr();
int ext3_prep_san_write(struct inode *inode, long *blocks,
Index: linux-2.6.9-full/fs/ext3/mballoc.c
===================================================================
---- linux-2.6.9-full.orig/fs/ext3/mballoc.c 2007-02-13 18:39:59.640066087 +0300
-+++ linux-2.6.9-full/fs/ext3/mballoc.c 2007-03-29 00:28:40.000000000 +0400
-@@ -0,0 +1,4342 @@
+--- linux-2.6.9-full.orig/fs/ext3/mballoc.c 2007-10-17 21:59:51.072534980 +0400
++++ linux-2.6.9-full/fs/ext3/mballoc.c 2007-10-17 23:09:22.000000000 +0400
+@@ -0,0 +1,4398 @@
+/*
+ * Copyright (c) 2003-2006, Cluster File Systems, Inc, info@clusterfs.com
+ * Written by Alex Tomas <alex@clusterfs.com>
+ ext2_set_bit(bit, addr);
+}
+
-+static inline void mb_set_bit_atomic(int bit, void *addr)
++static inline void mb_set_bit_atomic(spinlock_t *lock, int bit, void *addr)
+{
+ mb_correct_addr_and_bit(bit,addr);
-+ ext2_set_bit_atomic(NULL, bit, addr);
++ ext2_set_bit_atomic(lock, bit, addr);
+}
+
+static inline void mb_clear_bit(int bit, void *addr)
+ ext2_clear_bit(bit, addr);
+}
+
-+static inline void mb_clear_bit_atomic(int bit, void *addr)
++static inline void mb_clear_bit_atomic(spinlock_t *lock, int bit, void *addr)
+{
+ mb_correct_addr_and_bit(bit,addr);
-+ ext2_clear_bit_atomic(NULL, bit, addr);
++ ext2_clear_bit_atomic(lock, bit, addr);
+}
+
+static inline int mb_find_next_zero_bit(void *addr, int max, int start)
+}
+
+#ifdef DOUBLE_CHECK
-+void mb_free_blocks_double(struct ext3_buddy *e3b, int first, int count)
++void mb_free_blocks_double(struct inode *inode, struct ext3_buddy *e3b,
++ int first, int count)
+{
+ int i;
++ struct super_block *sb = e3b->bd_sb;
++
+ if (unlikely(e3b->bd_info->bb_bitmap == NULL))
+ return;
-+ BUG_ON(!ext3_is_group_locked(e3b->bd_sb, e3b->bd_group));
++ BUG_ON(!ext3_is_group_locked(sb, e3b->bd_group));
+ for (i = 0; i < count; i++) {
-+ BUG_ON(!mb_test_bit(first + i, e3b->bd_info->bb_bitmap));
++ if (!mb_test_bit(first + i, e3b->bd_info->bb_bitmap)) {
++ unsigned long blocknr;
++ blocknr = e3b->bd_group * EXT3_BLOCKS_PER_GROUP(sb);
++ blocknr += first + i;
++ blocknr +=
++ le32_to_cpu(EXT3_SB(sb)->s_es->s_first_data_block);
++
++ ext3_error(sb, __FUNCTION__, "double-free of inode"
++ " %lu's block %lu(bit %u in group %u)\n",
++ inode ? inode->i_ino : 0, blocknr,
++ first + i, e3b->bd_group);
++ }
+ mb_clear_bit(first + i, e3b->bd_info->bb_bitmap);
+ }
+}
+}
+
+#else
-+#define mb_free_blocks_double(a,b,c)
++#define mb_free_blocks_double(a,b,c,d)
+#define mb_mark_used_double(a,b,c)
+#define mb_cmp_bitmaps(a,b)
+#endif
+ return 0;
+}
+
-+static inline void mb_clear_bits(void *bm, int cur, int len)
++static inline void mb_clear_bits(spinlock_t *lock, void *bm, int cur, int len)
+{
+ __u32 *addr;
+
+ cur += 32;
+ continue;
+ }
-+ mb_clear_bit_atomic(cur, bm);
++ mb_clear_bit_atomic(lock, cur, bm);
+ cur++;
+ }
+}
+
-+static inline void mb_set_bits(void *bm, int cur, int len)
++static inline void mb_set_bits(spinlock_t *lock, void *bm, int cur, int len)
+{
+ __u32 *addr;
+
+ cur += 32;
+ continue;
+ }
-+ mb_set_bit_atomic(cur, bm);
++ mb_set_bit_atomic(lock, cur, bm);
+ cur++;
+ }
+}
+
-+static int mb_free_blocks(struct ext3_buddy *e3b, int first, int count)
++static int mb_free_blocks(struct inode *inode, struct ext3_buddy *e3b,
++ int first, int count)
+{
+ int block = 0, max = 0, order;
+ void *buddy, *buddy2;
++ struct super_block *sb = e3b->bd_sb;
+
-+ BUG_ON(first + count > (e3b->bd_sb->s_blocksize << 3));
-+ BUG_ON(!ext3_is_group_locked(e3b->bd_sb, e3b->bd_group));
++ BUG_ON(first + count > (sb->s_blocksize << 3));
++ BUG_ON(!ext3_is_group_locked(sb, e3b->bd_group));
+ mb_check_buddy(e3b);
-+ mb_free_blocks_double(e3b, first, count);
++ mb_free_blocks_double(inode, e3b, first, count);
+
+ e3b->bd_info->bb_free += count;
+ if (first < e3b->bd_info->bb_first_free)
+ /* let's maintain fragments counter */
+ if (first != 0)
+ block = !mb_test_bit(first - 1, EXT3_MB_BITMAP(e3b));
-+ if (first + count < EXT3_SB(e3b->bd_sb)->s_mb_maxs[0])
++ if (first + count < EXT3_SB(sb)->s_mb_maxs[0])
+ max = !mb_test_bit(first + count, EXT3_MB_BITMAP(e3b));
+ if (block && max)
+ e3b->bd_info->bb_fragments--;
+ block = first++;
+ order = 0;
+
-+ BUG_ON(!mb_test_bit(block, EXT3_MB_BITMAP(e3b)));
++ if (!mb_test_bit(block, EXT3_MB_BITMAP(e3b))) {
++ unsigned long blocknr;
++ blocknr = e3b->bd_group * EXT3_BLOCKS_PER_GROUP(sb);
++ blocknr += block;
++ blocknr +=
++ le32_to_cpu(EXT3_SB(sb)->s_es->s_first_data_block);
++
++ ext3_error(sb, __FUNCTION__, "double-free of inode"
++ " %lu's block %lu(bit %u in group %u)\n",
++ inode ? inode->i_ino : 0, blocknr, block,
++ e3b->bd_group);
++ }
+ mb_clear_bit(block, EXT3_MB_BITMAP(e3b));
+ e3b->bd_info->bb_counters[order]++;
+
+ e3b->bd_info->bb_counters[ord]++;
+ }
+
-+ mb_set_bits(EXT3_MB_BITMAP(e3b), ex->fe_start, len0);
++ mb_set_bits(sb_bgl_lock(EXT3_SB(e3b->bd_sb), ex->fe_group),
++ EXT3_MB_BITMAP(e3b), ex->fe_start, len0);
+ mb_check_buddy(e3b);
+
+ return ret;
+ struct ext3_free_extent *gex = &ac->ac_g_ex;
+
+ BUG_ON(ex->fe_len <= 0);
-+ BUG_ON(ex->fe_len >= (1 << ac->ac_sb->s_blocksize_bits) * 8);
-+ BUG_ON(ex->fe_start >= (1 << ac->ac_sb->s_blocksize_bits) * 8);
++ BUG_ON(ex->fe_len >= EXT3_BLOCKS_PER_GROUP(ac->ac_sb));
++ BUG_ON(ex->fe_start >= EXT3_BLOCKS_PER_GROUP(ac->ac_sb));
+ BUG_ON(ac->ac_status != AC_STATUS_CONTINUE);
+
+ ac->ac_found++;
+ i = e3b->bd_info->bb_first_free;
+
+ while (free && ac->ac_status == AC_STATUS_CONTINUE) {
-+ i = mb_find_next_zero_bit(bitmap, sb->s_blocksize * 8, i);
-+ if (i >= sb->s_blocksize * 8) {
++ i = mb_find_next_zero_bit(bitmap, EXT3_BLOCKS_PER_GROUP(sb), i);
++ if (i >= EXT3_BLOCKS_PER_GROUP(sb)) {
+ BUG_ON(free != 0);
+ break;
+ }
+ i = (i - le32_to_cpu(sbi->s_es->s_first_data_block))
+ % EXT3_BLOCKS_PER_GROUP(sb);
+
-+ while (i < sb->s_blocksize * 8) {
++ while (i < EXT3_BLOCKS_PER_GROUP(sb)) {
+ if (!mb_test_bit(i, bitmap)) {
+ max = mb_find_extent(e3b, 0, i, sbi->s_stripe, &ex);
+ if (max >= sbi->s_stripe) {
+ spin_lock_init(&sbi->s_mb_history_lock);
+ i = sbi->s_mb_history_max * sizeof(struct ext3_mb_history);
+ sbi->s_mb_history = kmalloc(i, GFP_KERNEL);
-+ memset(sbi->s_mb_history, 0, i);
++ if (likely(sbi->s_mb_history != NULL))
++ memset(sbi->s_mb_history, 0, i);
+ /* if we can't allocate history, then we simple won't use it */
+}
+
+ struct ext3_sb_info *sbi = EXT3_SB(ac->ac_sb);
+ struct ext3_mb_history h;
+
-+ if (likely(sbi->s_mb_history == NULL))
++ if (unlikely(sbi->s_mb_history == NULL))
+ return;
+
+ if (!(ac->ac_op & sbi->s_mb_history_filter))
+ sbi->s_mb_order2_reqs = MB_DEFAULT_ORDER2_REQS;
+ sbi->s_mb_history_filter = EXT3_MB_HISTORY_DEFAULT;
+
-+ i = sizeof(struct ext3_locality_group) * NR_CPUS;
++ i = sizeof(struct ext3_locality_group) * num_possible_cpus();
+ sbi->s_locality_groups = kmalloc(i, GFP_NOFS);
+ if (sbi->s_locality_groups == NULL) {
+ clear_opt(sbi->s_mount_opt, MBALLOC);
+ kfree(sbi->s_mb_maxs);
+ return -ENOMEM;
+ }
-+ for (i = 0; i < NR_CPUS; i++) {
++ for (i = 0; i < num_possible_cpus(); i++) {
+ struct ext3_locality_group *lg;
+ lg = &sbi->s_locality_groups[i];
+ sema_init(&lg->lg_sem, 1);
+ ext3_lock_group(sb, md->group);
+ for (i = 0; i < md->num; i++) {
+ mb_debug(" %u", md->blocks[i]);
-+ err = mb_free_blocks(&e3b, md->blocks[i], 1);
++ err = mb_free_blocks(NULL, &e3b, md->blocks[i], 1);
+ BUG_ON(err != 0);
+ }
+ mb_debug("\n");
+ struct ext3_sb_info *sbi;
+ struct super_block *sb;
+ sector_t block;
-+ int len, err;
++ int err;
+
+ BUG_ON(ac->ac_status != AC_STATUS_FOUND);
+ BUG_ON(ac->ac_b_ex.fe_len <= 0);
+ ext3_debug("using block group %d(%d)\n", ac->ac_b_group.group,
+ gdp->bg_free_blocks_count);
+
-+ /* time to check quota, we can't do this before because
-+ * having quota spent on preallocated-unused-yet blocks
-+ * would be wrong */
-+ len = ac->ac_b_ex.fe_len;
-+ while (len && DQUOT_ALLOC_BLOCK(ac->ac_inode, len)) len--;
-+ if (ac->ac_b_ex.fe_len != len) {
-+ /* some blocks can't be allocated due to quota
-+ * we have to return them back */
-+ BUG();
-+ }
-+ err = -EDQUOT;
-+ if (len == 0)
-+ goto out_err;
-+
+ err = -EIO;
+ bitmap_bh = read_block_bitmap(sb, ac->ac_b_ex.fe_group);
+ if (!bitmap_bh)
+ }
+ }
+#endif
-+ mb_set_bits(bitmap_bh->b_data, ac->ac_b_ex.fe_start, ac->ac_b_ex.fe_len);
++ mb_set_bits(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group), bitmap_bh->b_data,
++ ac->ac_b_ex.fe_start, ac->ac_b_ex.fe_len);
+
+ spin_lock(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group));
+ gdp->bg_free_blocks_count =
+ struct ext3_inode_info *ei = EXT3_I(ac->ac_inode);
+ loff_t start, end, size, orig_size, orig_start;
+ struct list_head *cur;
-+ int bsbits;
++ int bsbits, max;
+
+ /* do normalize only data requests, metadata requests
+ do not need preallocation */
+ if (size < i_size_read(ac->ac_inode))
+ size = i_size_read(ac->ac_inode);
+
++ /* max available blocks in a free group */
++ max = EXT3_BLOCKS_PER_GROUP(ac->ac_sb) - 1 - 1
++ - EXT3_SB(ac->ac_sb)->s_itb_per_group;
++
++#define NRL_CHECK_SIZE(req,size,max,bits) \
++ (req <= (size) || max <= ((size) >> bits))
++
+ /* first, try to predict filesize */
+ /* XXX: should this table be tunable? */
+ start = 0;
+ size = 512 * 1024;
+ } else if (size <= 1024 * 1024) {
+ size = 1024 * 1024;
-+ } else if (size < 4 * 1024 * 1024) {
++ } else if (NRL_CHECK_SIZE(size, 4 * 1024 * 1024, max, bsbits)) {
+ start = ac->ac_o_ex.fe_logical << bsbits;
+ start = (start / (1024 * 1024)) * (1024 * 1024);
+ size = 1024 * 1024;
-+ } else if (size < 8 * 1024 * 1024) {
++ } else if (NRL_CHECK_SIZE(size, 8 * 1024 * 1024, max, bsbits)) {
+ start = ac->ac_o_ex.fe_logical << bsbits;
+ start = (start / (4 * (1024 * 1024))) * 4 * (1024 * 1024);
+ size = 4 * 1024 * 1024;
-+ } else if (ac->ac_o_ex.fe_len < ((8 << 20) >> bsbits)) {
++ } else if(NRL_CHECK_SIZE(ac->ac_o_ex.fe_len,(8<<20)>>bsbits,max,bsbits)){
+ start = ac->ac_o_ex.fe_logical;
+ start = start << bsbits;
+ start = (start / (8 * (1024 * 1024))) * 8 * (1024 * 1024);
+ unsigned long pa_end;
+
+ pa = list_entry(cur, struct ext3_prealloc_space, pa_inode_list);
++
++ if (pa->pa_deleted)
++ continue;
++ spin_lock(&pa->pa_lock);
++ if (pa->pa_deleted) {
++ spin_unlock(&pa->pa_lock);
++ continue;
++ }
++
+ pa_end = pa->pa_lstart + pa->pa_len;
+
+ /* PA must not overlap original request */
+ ac->ac_o_ex.fe_logical < pa->pa_lstart));
+
+ /* skip PA normalized request doesn't overlap with */
-+ if (pa->pa_lstart >= end)
++ if (pa->pa_lstart >= end) {
++ spin_unlock(&pa->pa_lock);
+ continue;
-+ if (pa_end <= start)
++ }
++ if (pa_end <= start) {
++ spin_unlock(&pa->pa_lock);
+ continue;
++ }
+ BUG_ON(pa->pa_lstart <= start && pa_end >= end);
+
+ if (pa_end <= ac->ac_o_ex.fe_logical) {
+ BUG_ON(pa->pa_lstart > end);
+ end = pa->pa_lstart;
+ }
++ spin_unlock(&pa->pa_lock);
+ }
+ rcu_read_unlock();
+ size = end - start;
+ struct ext3_prealloc_space *pa;
+ unsigned long pa_end;
+ pa = list_entry(cur, struct ext3_prealloc_space, pa_inode_list);
-+ pa_end = pa->pa_lstart + pa->pa_len;
-+ BUG_ON(!(start >= pa_end || end <= pa->pa_lstart));
++ spin_lock(&pa->pa_lock);
++ if (pa->pa_deleted == 0) {
++ pa_end = pa->pa_lstart + pa->pa_len;
++ BUG_ON(!(start >= pa_end || end <= pa->pa_lstart));
++ }
++ spin_unlock(&pa->pa_lock);
+ }
+ rcu_read_unlock();
+
+ ext3_get_group_no_and_offset(sb, pa->pa_pstart, &groupnr, &start);
+ len = pa->pa_len;
+ spin_unlock(&pa->pa_lock);
-+ BUG_ON(groupnr != group);
-+ mb_set_bits(bitmap, start, len);
++ if (unlikely(len == 0))
++ continue;
++ BUG_ON(groupnr != group && len != 0);
++ mb_set_bits(sb_bgl_lock(EXT3_SB(sb), group), bitmap, start,len);
+ preallocated += len;
+ count++;
+ }
+ BUG_ON(ac->ac_status != AC_STATUS_FOUND);
+ BUG_ON(!S_ISREG(ac->ac_inode->i_mode));
+
-+ pa = kmem_cache_alloc(ext3_pspace_cachep, SLAB_NOFS);
++ pa = kmem_cache_alloc(ext3_pspace_cachep, GFP_NOFS);
+ if (pa == NULL)
+ return -ENOMEM;
+
+ BUG_ON(!S_ISREG(ac->ac_inode->i_mode));
+
+ BUG_ON(ext3_pspace_cachep == NULL);
-+ pa = kmem_cache_alloc(ext3_pspace_cachep, SLAB_NOFS);
++ pa = kmem_cache_alloc(ext3_pspace_cachep, GFP_NOFS);
+ if (pa == NULL)
+ return -ENOMEM;
+
+
+ BUG_ON(pa->pa_deleted == 0);
+ ext3_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit);
-+ BUG_ON(group != e3b->bd_group);
++ BUG_ON(group != e3b->bd_group && pa->pa_len != 0);
+ end = bit + pa->pa_len;
+
+ ac.ac_sb = sb;
+ ac.ac_b_ex.fe_logical = 0;
+ ext3_mb_store_history(&ac);
+
-+ mb_free_blocks(e3b, bit, next - bit);
++ mb_free_blocks(pa->pa_inode, e3b, bit, next - bit);
+ bit = next + 1;
+ }
+ if (free != pa->pa_free) {
+
+ BUG_ON(pa->pa_deleted == 0);
+ ext3_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit);
-+ BUG_ON(group != e3b->bd_group);
-+ mb_free_blocks(e3b, bit, pa->pa_len);
++ BUG_ON(group != e3b->bd_group && pa->pa_len != 0);
++ mb_free_blocks(pa->pa_inode, e3b, bit, pa->pa_len);
+ atomic_add(pa->pa_len, &EXT3_SB(sb)->s_mb_discarded);
+
+ ac.ac_sb = sb;
+ spin_lock(&pa->pa_lock);
+ if (atomic_read(&pa->pa_count)) {
+ spin_unlock(&pa->pa_lock);
-+ printk("uh! busy PA\n");
-+ dump_stack();
+ busy = 1;
+ continue;
+ }
+ }
+
+ /* if we still need more blocks and some PAs were used, try again */
-+ if (free < needed && busy)
++ if (free < needed && busy) {
++ ext3_unlock_group(sb, group);
+ goto repeat;
++ }
+
+ /* found anything to free? */
+ if (list_empty(&list)) {
+ * use preallocation while we're discarding it */
+ spin_unlock(&pa->pa_lock);
+ spin_unlock(&ei->i_prealloc_lock);
-+ printk("uh-oh! used pa while discarding\n");
-+ dump_stack();
+ current->state = TASK_UNINTERRUPTIBLE;
+ schedule_timeout(HZ);
+ goto repeat;
+ * add a flag to force wait only in case
+ * of ->clear_inode(), but not in case of
+ * regular truncate */
-+ printk("uh-oh! some one just deleted it\n");
-+ dump_stack();
+ current->state = TASK_UNINTERRUPTIBLE;
+ schedule_timeout(HZ);
+ goto repeat;
+ BUG_ON(err != 0); /* error handling here */
+
+ bitmap_bh = read_block_bitmap(sb, group);
-+ if (bitmap_bh == NULL) {
-+ /* error handling here */
-+ ext3_mb_release_desc(&e3b);
-+ BUG_ON(bitmap_bh == NULL);
-+ }
+
+ ext3_lock_group(sb, group);
+ list_del_rcu(&pa->pa_group_list);
-+ ext3_mb_release_inode_pa(&e3b, bitmap_bh, pa);
++
++ /* can be NULL due to IO error, at worst
++ * we leave some free blocks unavailable
++ * do not go RO - no need for */
++ if (bitmap_bh != NULL)
++ ext3_mb_release_inode_pa(&e3b, bitmap_bh, pa);
+ ext3_unlock_group(sb, group);
+
+ ext3_mb_release_desc(&e3b);
+ struct ext3_allocation_context ac;
+ struct ext3_sb_info *sbi;
+ struct super_block *sb;
-+ unsigned long block;
-+ int err, freed;
++ unsigned long block = 0;
++ int freed, inquota;
+
+ sb = ar->inode->i_sb;
+ sbi = EXT3_SB(sb);
+ printk(KERN_ERR "EXT3-fs: multiblock request with "
+ "mballoc disabled!\n");
+ ar->len = 1;
-+ err = ext3_new_block_old(handle, ar->inode, ar->goal, errp);
-+ return err;
++ block = ext3_new_block_old(handle, ar->inode, ar->goal, errp);
++ return block;
++ }
++
++ while (ar->len && DQUOT_ALLOC_BLOCK(ar->inode, ar->len)) {
++ ar->flags |= EXT3_MB_HINT_NOPREALLOC;
++ ar->len--;
++ }
++ if (ar->len == 0) {
++ *errp = -EDQUOT;
++ return 0;
+ }
++ inquota = ar->len;
+
+ ext3_mb_poll_new_transaction(sb, handle);
+
-+ if ((err = ext3_mb_initialize_context(&ac, ar)))
-+ return err;
++ if ((*errp = ext3_mb_initialize_context(&ac, ar))) {
++ ar->len = 0;
++ goto out;
++ }
+
+ ac.ac_op = EXT3_MB_HISTORY_PREALLOC;
+ if (!ext3_mb_use_preallocated(&ac)) {
+ goto repeat;
+ *errp = -ENOSPC;
+ ac.ac_b_ex.fe_len = 0;
-+ block = 0;
++ ar->len = 0;
+ ext3_mb_show_ac(&ac);
+ }
+
+ ext3_mb_release_context(&ac);
+
++out:
++ if (ar->len < inquota)
++ DQUOT_FREE_BLOCK(ar->inode, inquota - ar->len);
++
+ return block;
+}
+EXPORT_SYMBOL(ext3_mb_new_blocks);
+ BUG_ON(!mb_test_bit(bit + i, bitmap_bh->b_data));
+ }
+#endif
-+ mb_clear_bits(bitmap_bh->b_data, bit, count);
++ mb_clear_bits(sb_bgl_lock(sbi, block_group), bitmap_bh->b_data, bit,
++ count);
+
+ /* We dirtied the bitmap block */
+ BUFFER_TRACE(bitmap_bh, "dirtied bitmap block");
+ ext3_mb_free_metadata(handle, &e3b, block_group, bit, count);
+ } else {
+ ext3_lock_group(sb, block_group);
-+ err = mb_free_blocks(&e3b, bit, count);
++ err = mb_free_blocks(inode, &e3b, bit, count);
+ ext3_mb_return_to_preallocation(inode, &e3b, block, count);
+ ext3_unlock_group(sb, block_group);
+ BUG_ON(err != 0);