summaryrefslogtreecommitdiffstats
path: root/drivers/md
diff options
context:
space:
mode:
authorNeilBrown <neilb@suse.de>2009-01-09 08:31:08 +1100
committerNeilBrown <neilb@suse.de>2009-01-09 08:31:08 +1100
commit8b76539823d71576927e3eb08b395eb6620f628d (patch)
tree37f9c4da6c861a69b4695ea7818eab2cf6ce95f4 /drivers/md
parentcd2ac9321c26dc7a76455cd2a4df89123fa2b73e (diff)
downloadlinux-stable-8b76539823d71576927e3eb08b395eb6620f628d.tar.gz
linux-stable-8b76539823d71576927e3eb08b395eb6620f628d.tar.bz2
linux-stable-8b76539823d71576927e3eb08b395eb6620f628d.zip
md: move allocation of ->queue from mddev_find to md_probe
It is more balanced to just do simple initialisation in mddev_find, which allocates and links a new md device, and leave all the more sophisticated allocation to md_probe (which calls mddev_find). md_probe already allocated the gendisk. It should allocate the queue too. Signed-off-by: NeilBrown <neilb@suse.de>
Diffstat (limited to 'drivers/md')
-rw-r--r--drivers/md/md.c28
1 files changed, 17 insertions, 11 deletions
diff --git a/drivers/md/md.c b/drivers/md/md.c
index 1f770c16d435..da838cc32cc9 100644
--- a/drivers/md/md.c
+++ b/drivers/md/md.c
@@ -221,7 +221,9 @@ static void mddev_put(mddev_t *mddev)
if (!mddev->raid_disks && list_empty(&mddev->disks)) {
list_del(&mddev->all_mddevs);
spin_unlock(&all_mddevs_lock);
- blk_cleanup_queue(mddev->queue);
+ if (mddev->queue)
+ blk_cleanup_queue(mddev->queue);
+ mddev->queue = NULL;
if (mddev->sysfs_state)
sysfs_put(mddev->sysfs_state);
mddev->sysfs_state = NULL;
@@ -275,16 +277,6 @@ static mddev_t * mddev_find(dev_t unit)
new->resync_max = MaxSector;
new->level = LEVEL_NONE;
- new->queue = blk_alloc_queue(GFP_KERNEL);
- if (!new->queue) {
- kfree(new);
- return NULL;
- }
- /* Can be unlocked because the queue is new: no concurrency */
- queue_flag_set_unlocked(QUEUE_FLAG_CLUSTER, new->queue);
-
- blk_queue_make_request(new->queue, md_fail_request);
-
goto retry;
}
@@ -3493,9 +3485,23 @@ static struct kobject *md_probe(dev_t dev, int *part, void *data)
mddev_put(mddev);
return NULL;
}
+
+ mddev->queue = blk_alloc_queue(GFP_KERNEL);
+ if (!mddev->queue) {
+ mutex_unlock(&disks_mutex);
+ mddev_put(mddev);
+ return NULL;
+ }
+ /* Can be unlocked because the queue is new: no concurrency */
+ queue_flag_set_unlocked(QUEUE_FLAG_CLUSTER, mddev->queue);
+
+ blk_queue_make_request(mddev->queue, md_fail_request);
+
disk = alloc_disk(1 << shift);
if (!disk) {
mutex_unlock(&disks_mutex);
+ blk_cleanup_queue(mddev->queue);
+ mddev->queue = NULL;
mddev_put(mddev);
return NULL;
}