[dm-devel] [PATCH v2 1/3] dm: support for non power of 2 target IO sizes
Mike Snitzer
snitzer at redhat.com
Sat May 5 03:57:08 UTC 2012
Update DM core's max_io_len() to use sector_div() to support a target's
use of a non power of 2 blocksize.
Rename split_io to max_io_len in the dm_target structure and change data
type from sector_t to uint32_t. Introduce dm_set_target_max_io_len() to
check that each target's max_io_len does not exceed UINT_MAX.
Add do_div wrappers (dm_do_div and dm_do_mod) that don't modify the
dividend; these will be used by thin-pool and striped targets.
Signed-off-by: Mike Snitzer <snitzer at redhat.com>
---
drivers/md/dm-raid.c | 12 +++++-------
drivers/md/dm-raid1.c | 4 +++-
drivers/md/dm-snap.c | 19 +++++++++++--------
drivers/md/dm-stripe.c | 4 +++-
drivers/md/dm-thin.c | 4 +++-
drivers/md/dm.c | 28 +++++++++++++++++++++-------
include/linux/device-mapper.h | 20 ++++++++++++++++++--
7 files changed, 64 insertions(+), 27 deletions(-)
Index: linux/drivers/md/dm-raid.c
===================================================================
--- linux.orig/drivers/md/dm-raid.c
+++ linux/drivers/md/dm-raid.c
@@ -356,6 +356,7 @@ static int parse_raid_params(struct raid
{
unsigned i, rebuild_cnt = 0;
unsigned long value, region_size = 0;
+ sector_t max_io_len;
char *key;
/*
@@ -525,14 +526,11 @@ static int parse_raid_params(struct raid
return -EINVAL;
if (rs->md.chunk_sectors)
- rs->ti->split_io = rs->md.chunk_sectors;
+ max_io_len = rs->md.chunk_sectors;
else
- rs->ti->split_io = region_size;
-
- if (rs->md.chunk_sectors)
- rs->ti->split_io = rs->md.chunk_sectors;
- else
- rs->ti->split_io = region_size;
+ max_io_len = region_size;
+ if (dm_set_target_max_io_len(rs->ti, max_io_len) < 0)
+ return -EINVAL;
/* Assume there are no metadata devices until the drives are parsed */
rs->md.persistent = 0;
Index: linux/drivers/md/dm-raid1.c
===================================================================
--- linux.orig/drivers/md/dm-raid1.c
+++ linux/drivers/md/dm-raid1.c
@@ -1081,7 +1081,9 @@ static int mirror_ctr(struct dm_target *
}
ti->private = ms;
- ti->split_io = dm_rh_get_region_size(ms->rh);
+ r = dm_set_target_max_io_len(ti, dm_rh_get_region_size(ms->rh));
+ if (r < 0)
+ goto err_free_context;
ti->num_flush_requests = 1;
ti->num_discard_requests = 1;
Index: linux/drivers/md/dm-snap.c
===================================================================
--- linux.orig/drivers/md/dm-snap.c
+++ linux/drivers/md/dm-snap.c
@@ -1172,7 +1172,10 @@ static int snapshot_ctr(struct dm_target
ti->error = "Chunk size not set";
goto bad_read_metadata;
}
- ti->split_io = s->store->chunk_size;
+
+ r = dm_set_target_max_io_len(ti, s->store->chunk_size);
+ if (r < 0)
+ goto bad_read_metadata;
return 0;
@@ -1239,7 +1242,7 @@ static void __handover_exceptions(struct
snap_dest->store->snap = snap_dest;
snap_src->store->snap = snap_src;
- snap_dest->ti->split_io = snap_dest->store->chunk_size;
+ snap_dest->ti->max_io_len = snap_dest->store->chunk_size;
snap_dest->valid = snap_src->valid;
/*
@@ -1838,9 +1841,9 @@ static void snapshot_merge_resume(struct
snapshot_resume(ti);
/*
- * snapshot-merge acts as an origin, so set ti->split_io
+ * snapshot-merge acts as an origin, so set ti->max_io_len
*/
- ti->split_io = get_origin_minimum_chunksize(s->origin->bdev);
+ ti->max_io_len = get_origin_minimum_chunksize(s->origin->bdev);
start_merge(s);
}
@@ -2073,12 +2076,12 @@ static int origin_write_extent(struct dm
struct origin *o;
/*
- * The origin's __minimum_chunk_size() got stored in split_io
+ * The origin's __minimum_chunk_size() got stored in max_io_len
* by snapshot_merge_resume().
*/
down_read(&_origins_lock);
o = __lookup_origin(merging_snap->origin->bdev);
- for (n = 0; n < size; n += merging_snap->ti->split_io)
+ for (n = 0; n < size; n += merging_snap->ti->max_io_len)
if (__origin_write(&o->snapshots, sector + n, NULL) ==
DM_MAPIO_SUBMITTED)
must_wait = 1;
@@ -2138,14 +2141,14 @@ static int origin_map(struct dm_target *
}
/*
- * Set the target "split_io" field to the minimum of all the snapshots'
+ * Set the target "max_io_len" field to the minimum of all the snapshots'
* chunk sizes.
*/
static void origin_resume(struct dm_target *ti)
{
struct dm_dev *dev = ti->private;
- ti->split_io = get_origin_minimum_chunksize(dev->bdev);
+ ti->max_io_len = get_origin_minimum_chunksize(dev->bdev);
}
static int origin_status(struct dm_target *ti, status_type_t type, char *result,
Index: linux/drivers/md/dm-stripe.c
===================================================================
--- linux.orig/drivers/md/dm-stripe.c
+++ linux/drivers/md/dm-stripe.c
@@ -172,7 +172,9 @@ static int stripe_ctr(struct dm_target *
sc->stripes_mask = ((sector_t) stripes) - 1;
}
- ti->split_io = chunk_size;
+ r = dm_set_target_max_io_len(ti, chunk_size);
+ if (r < 0)
+ return r;
ti->num_flush_requests = stripes;
ti->num_discard_requests = stripes;
Index: linux/drivers/md/dm-thin.c
===================================================================
--- linux.orig/drivers/md/dm-thin.c
+++ linux/drivers/md/dm-thin.c
@@ -2575,7 +2575,9 @@ static int thin_ctr(struct dm_target *ti
goto bad_thin_open;
}
- ti->split_io = tc->pool->sectors_per_block;
+ r = dm_set_target_max_io_len(ti, tc->pool->sectors_per_block);
+ if (r < 0)
+ goto bad_thin_open;
ti->num_flush_requests = 1;
/* In case the pool supports discards, pass them on. */
Index: linux/drivers/md/dm.c
===================================================================
--- linux.orig/drivers/md/dm.c
+++ linux/drivers/md/dm.c
@@ -970,20 +970,34 @@ static sector_t max_io_len(sector_t sect
sector_t len = max_io_len_target_boundary(sector, ti);
/*
- * Does the target need to split even further ?
+ * Does the target need to split even further?
*/
- if (ti->split_io) {
- sector_t boundary;
+ if (ti->max_io_len) {
sector_t offset = dm_target_offset(ti, sector);
- boundary = ((offset + ti->split_io) & ~(ti->split_io - 1))
- - offset;
- if (len > boundary)
- len = boundary;
+ sector_t boundary = offset + ti->max_io_len;
+ sector_t max_len = ti->max_io_len - sector_div(boundary, ti->max_io_len);
+
+ if (len > max_len)
+ len = max_len;
}
return len;
}
+int dm_set_target_max_io_len(struct dm_target *ti, sector_t len)
+{
+ if (len > UINT_MAX) {
+ DMERR("Specified maximum size of target IO (%llu) exceeds limit (%u)",
+ (unsigned long long)len, UINT_MAX);
+ ti->error = "Maximum size of target IO is too large";
+ return -EINVAL;
+ }
+
+ ti->max_io_len = (uint32_t) len;
+ return 0;
+}
+EXPORT_SYMBOL_GPL(dm_set_target_max_io_len);
+
static void __map_bio(struct dm_target *ti, struct bio *clone,
struct dm_target_io *tio)
{
Index: linux/include/linux/device-mapper.h
===================================================================
--- linux.orig/include/linux/device-mapper.h
+++ linux/include/linux/device-mapper.h
@@ -188,8 +188,8 @@ struct dm_target {
sector_t begin;
sector_t len;
- /* Always a power of 2 */
- sector_t split_io;
+ /* If non-zero, maximum size of I/O submitted to a target. */
+ uint32_t max_io_len;
/*
* A number of zero-length barrier requests that will be submitted
@@ -503,6 +503,20 @@ static inline unsigned long to_bytes(sec
return (n << SECTOR_SHIFT);
}
+/*
+ * do_div wrappers that don't modify the dividend
+ */
+static inline sector_t dm_do_div(sector_t a, uint32_t b)
+{
+ do_div(a, b);
+ return a;
+}
+
+static inline uint32_t dm_do_mod(sector_t a, uint32_t b)
+{
+ return do_div(a, b);
+}
+
/*-----------------------------------------------------------------
* Helper for block layer and dm core operations
*---------------------------------------------------------------*/
@@ -511,4 +525,6 @@ void dm_requeue_unmapped_request(struct
void dm_kill_unmapped_request(struct request *rq, int error);
int dm_underlying_device_busy(struct request_queue *q);
+int __must_check dm_set_target_max_io_len(struct dm_target *ti, sector_t len);
+
#endif /* _LINUX_DEVICE_MAPPER_H */
More information about the dm-devel
mailing list