Give each device its own io mempool to avoid a potential
deadlock with stacked devices. [HM + EJT]
--- diff/drivers/md/dm.c 2002-11-29 09:45:55.000000000 +0000
+++ source/drivers/md/dm.c 2002-11-29 15:06:54.000000000 +0000
@@ -62,11 +62,15 @@
* The current mapping.
*/
struct dm_table *map;
+
+ /*
+ * io objects are allocated from here.
+ */
+ mempool_t *io_pool;
};
#define MIN_IOS 256
static kmem_cache_t *_io_cache;
-static mempool_t *_io_pool;
/* block device arrays */
static int _block_size[MAX_DEVICES];
@@ -89,18 +93,10 @@
if (!_io_cache)
return -ENOMEM;
- _io_pool = mempool_create(MIN_IOS, mempool_alloc_slab,
- mempool_free_slab, _io_cache);
- if (!_io_pool) {
- kmem_cache_destroy(_io_cache);
- return -ENOMEM;
- }
-
_major = major;
r = register_blkdev(_major, _name, &dm_blk_dops);
if (r < 0) {
DMERR("register_blkdev failed");
- mempool_destroy(_io_pool);
kmem_cache_destroy(_io_cache);
return r;
}
@@ -121,7 +117,6 @@
static void local_exit(void)
{
- mempool_destroy(_io_pool);
kmem_cache_destroy(_io_cache);
if (unregister_blkdev(_major, _name) < 0)
@@ -209,14 +204,14 @@
return 0;
}
-static inline struct dm_io *alloc_io(void)
+static inline struct dm_io *alloc_io(struct mapped_device *md)
{
- return mempool_alloc(_io_pool, GFP_NOIO);
+ return mempool_alloc(md->io_pool, GFP_NOIO);
}
-static inline void free_io(struct dm_io *io)
+static inline void free_io(struct mapped_device *md, struct dm_io *io)
{
- mempool_free(io, _io_pool);
+ mempool_free(io, md->io_pool);
}
static inline struct deferred_io *alloc_deferred(void)
@@ -326,7 +321,7 @@
bh->b_end_io = io->end_io;
bh->b_private = io->context;
- free_io(io);
+ free_io(io->md, io);
bh->b_end_io(bh, uptodate);
}
@@ -345,7 +340,7 @@
if (!ti)
return -EINVAL;
- io = alloc_io();
+ io = alloc_io(md);
if (!io)
return -ENOMEM;
@@ -363,7 +358,7 @@
} else
/* we don't need to hook */
- free_io(io);
+ free_io(io->md, io);
return r;
}
@@ -609,6 +604,15 @@
DMWARN("allocating minor %d.", minor);
memset(md, 0, sizeof(*md));
+
+ md->io_pool = mempool_create(MIN_IOS, mempool_alloc_slab,
+ mempool_free_slab, _io_cache);
+ if (!md->io_pool) {
+ free_minor(minor);
+ kfree(md);
+ return NULL;
+ }
+
md->dev = mk_kdev(_major, minor);
init_rwsem(&md->lock);
atomic_set(&md->holders, 1);
@@ -621,6 +625,7 @@
static void free_dev(struct mapped_device *md)
{
free_minor(minor(md->dev));
+ mempool_destroy(md->io_pool);
kfree(md);
}