aboutsummaryrefslogtreecommitdiffstats
path: root/fs
diff options
context:
space:
mode:
authorSteve French <sfrench@us.ibm.com>2005-11-01 09:02:10 -0800
committerSteve French <sfrench@us.ibm.com>2005-11-01 09:02:10 -0800
commitef0eaa13362041f531f951b89362bfbc177e2bc0 (patch)
treee2d43099698aa238f9d108c67451871b0d4073f9 /fs
parentMerge with /pub/scm/linux/kernel/git/torvalds/linux-2.6.git (diff)
parentMerge git://git.kernel.org/pub/scm/linux/kernel/git/aia21/ntfs-2.6 (diff)
downloadlinux-dev-ef0eaa13362041f531f951b89362bfbc177e2bc0.tar.xz
linux-dev-ef0eaa13362041f531f951b89362bfbc177e2bc0.zip
Merge with /pub/scm/linux/kernel/git/torvalds/linux-2.6.git
Diffstat (limited to 'fs')
-rw-r--r--fs/fs-writeback.c2
-rw-r--r--fs/ntfs/file.c17
2 files changed, 10 insertions, 9 deletions
diff --git a/fs/fs-writeback.c b/fs/fs-writeback.c
index ffab4783ac64..c27f8d4098be 100644
--- a/fs/fs-writeback.c
+++ b/fs/fs-writeback.c
@@ -247,7 +247,7 @@ __writeback_single_inode(struct inode *inode, struct writeback_control *wbc)
wait_queue_head_t *wqh;
if (!atomic_read(&inode->i_count))
- WARN_ON(!(inode->i_state & I_WILL_FREE));
+ WARN_ON(!(inode->i_state & (I_WILL_FREE|I_FREEING)));
else
WARN_ON(inode->i_state & I_WILL_FREE);
diff --git a/fs/ntfs/file.c b/fs/ntfs/file.c
index cf3e6ced2d01..727533891813 100644
--- a/fs/ntfs/file.c
+++ b/fs/ntfs/file.c
@@ -668,10 +668,10 @@ map_buffer_cached:
* to, we need to read it in before the write,
* i.e. now.
*/
- if (!buffer_uptodate(bh) && ((bh_pos < pos &&
- bh_end > pos) ||
- (bh_end > end &&
- bh_end > end))) {
+ if (!buffer_uptodate(bh) && bh_pos < end &&
+ bh_end > pos &&
+ (bh_pos < pos ||
+ bh_end > end)) {
/*
* If the buffer is fully or partially
* within the initialized size, do an
@@ -784,10 +784,11 @@ retry_remap:
blocksize_bits);
cdelta = 0;
/*
- * If the number of remaining clusters in the
- * @pages is smaller or equal to the number of
- * cached clusters, unlock the runlist as the
- * map cache will be used from now on.
+ * If the number of remaining clusters touched
+ * by the write is smaller or equal to the
+ * number of cached clusters, unlock the
+ * runlist as the map cache will be used from
+ * now on.
*/
if (likely(vcn + vcn_len >= cend)) {
if (rl_write_locked) {