summaryrefslogtreecommitdiffstats
path: root/include/linux/backing-dev.h
diff options
context:
space:
mode:
authorJan Kara <jack@suse.cz>2014-04-03 14:46:23 -0700
committerLinus Torvalds <torvalds@linux-foundation.org>2014-04-03 16:20:49 -0700
commit5acda9d12dcf1ad0d9a5a2a7c646de3472fa7555 (patch)
tree5103d2779fbf2f1f07edb1ed51083d1303073ef9 /include/linux/backing-dev.h
parent6ca738d60c563d5c6cf6253ee4b8e76fa77b2b9e (diff)
downloadlinux-5acda9d12dcf1ad0d9a5a2a7c646de3472fa7555.tar.gz
linux-5acda9d12dcf1ad0d9a5a2a7c646de3472fa7555.tar.bz2
linux-5acda9d12dcf1ad0d9a5a2a7c646de3472fa7555.zip
bdi: avoid oops on device removal
After commit 839a8e8660b6 ("writeback: replace custom worker pool implementation with unbound workqueue") when device is removed while we are writing to it we crash in bdi_writeback_workfn() -> set_worker_desc() because bdi->dev is NULL. This can happen because even though bdi_unregister() cancels all pending flushing work, nothing really prevents new ones from being queued from balance_dirty_pages() or other places. Fix the problem by clearing BDI_registered bit in bdi_unregister() and checking it before scheduling of any flushing work. Fixes: 839a8e8660b6777e7fe4e80af1a048aebe2b5977 Reviewed-by: Tejun Heo <tj@kernel.org> Signed-off-by: Jan Kara <jack@suse.cz> Cc: Derek Basehore <dbasehore@chromium.org> Cc: Jens Axboe <axboe@kernel.dk> Cc: <stable@vger.kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'include/linux/backing-dev.h')
-rw-r--r--include/linux/backing-dev.h2
1 files changed, 1 insertions, 1 deletions
diff --git a/include/linux/backing-dev.h b/include/linux/backing-dev.h
index 24819001f5c8..e488e9459a93 100644
--- a/include/linux/backing-dev.h
+++ b/include/linux/backing-dev.h
@@ -95,7 +95,7 @@ struct backing_dev_info {
unsigned int max_ratio, max_prop_frac;
struct bdi_writeback wb; /* default writeback info for this bdi */
- spinlock_t wb_lock; /* protects work_list */
+ spinlock_t wb_lock; /* protects work_list & wb.dwork scheduling */
struct list_head work_list;