summaryrefslogtreecommitdiffstats
path: root/fs/btrfs
diff options
context:
space:
mode:
authorWu Fengguang <fengguang.wu@intel.com>2011-12-16 12:32:57 -0500
committerChris Mason <chris.mason@oracle.com>2011-12-16 12:32:57 -0500
commit142349f541d0bb6bc3e0d4563268105aada42b0b (patch)
treec86e13b5cf8b87b3b4f3c5218442e221ce17f67b /fs/btrfs
parentdc47ce90c3a822cd7c9e9339fe4d5f61dcb26b50 (diff)
downloadlinux-stable-142349f541d0bb6bc3e0d4563268105aada42b0b.tar.gz
linux-stable-142349f541d0bb6bc3e0d4563268105aada42b0b.tar.bz2
linux-stable-142349f541d0bb6bc3e0d4563268105aada42b0b.zip
btrfs: lower the dirty balance poll interval
Tests show that the original large intervals can easily make the dirty limit exceeded on 100 concurrent dd's. So adapt to as large as the next check point selected by the dirty throttling algorithm. Signed-off-by: Wu Fengguang <fengguang.wu@intel.com> Signed-off-by: Chris Mason <chris.mason@oracle.com>
Diffstat (limited to 'fs/btrfs')
-rw-r--r--fs/btrfs/file.c2
1 files changed, 2 insertions, 0 deletions
diff --git a/fs/btrfs/file.c b/fs/btrfs/file.c
index dafdfa059bf6..52305a885c3f 100644
--- a/fs/btrfs/file.c
+++ b/fs/btrfs/file.c
@@ -1167,6 +1167,8 @@ static noinline ssize_t __btrfs_buffered_write(struct file *file,
nrptrs = min((iov_iter_count(i) + PAGE_CACHE_SIZE - 1) /
PAGE_CACHE_SIZE, PAGE_CACHE_SIZE /
(sizeof(struct page *)));
+ nrptrs = min(nrptrs, current->nr_dirtied_pause - current->nr_dirtied);
+ nrptrs = max(nrptrs, 8);
pages = kmalloc(nrptrs * sizeof(struct page *), GFP_KERNEL);
if (!pages)
return -ENOMEM;