block copy: introduce "copy_boundary" limits There is no way to split copy requests, so the creator of the requests (the function blkdev_issue_copy) must make requests with proper size. Device mapper splits the requests at a boundary between targets or at a boundary specified by each target driver. We must make sure that the copy requets do not cross these boundaries. This patch introduces a new queue limit "copy_boundary", it is log2 of the boundary in sectors that the request must not cross. Device mapper will use this limit to propagate its requirements through the device stack. Signed-off-by: Mikulas Patocka --- block/blk-lib.c | 21 ++++++++++++++++++++- block/blk-settings.c | 17 +++++++++++++++++ block/blk-sysfs.c | 13 +++++++++++++ include/linux/blkdev.h | 1 + 4 files changed, 51 insertions(+), 1 deletion(-) Index: linux-4.7-rc3/block/blk-settings.c =================================================================== --- linux-4.7-rc3.orig/block/blk-settings.c 2016-06-13 17:19:43.000000000 +0200 +++ linux-4.7-rc3/block/blk-settings.c 2016-06-13 17:19:45.000000000 +0200 @@ -96,6 +96,7 @@ void blk_set_default_limits(struct queue lim->chunk_sectors = 0; lim->max_write_same_sectors = 0; lim->max_copy_sectors = 0; + lim->copy_boundary = 63; lim->max_discard_sectors = 0; lim->max_hw_discard_sectors = 0; lim->discard_granularity = 0; @@ -311,6 +312,18 @@ void blk_queue_max_copy_sectors(struct r EXPORT_SYMBOL(blk_queue_max_copy_sectors); /** + * blk_queue_copy_boundary - set a boundary for copy operations. No copy + * operation may cross the boundary + * @q: the request queue for the device + * @copy_boundary: log2 of the copy boundary in sectors + **/ +void blk_queue_copy_boundary(struct request_queue *q, + unsigned char copy_boundary) +{ + q->limits.copy_boundary = copy_boundary; +} + +/** * blk_queue_max_segments - set max hw segments for a request for this queue * @q: the request queue for the device * @max_segments: max number of segments @@ -552,6 +565,10 @@ int blk_stack_limits(struct queue_limits t->max_segment_size = min_not_zero(t->max_segment_size, b->max_segment_size); + t->copy_boundary = min(t->copy_boundary, b->copy_boundary); + if (start) + t->copy_boundary = min(t->copy_boundary, (unsigned char)__ffs64(start)); + t->misaligned |= b->misaligned; alignment = queue_limit_alignment_offset(b, start); Index: linux-4.7-rc3/block/blk-sysfs.c =================================================================== --- linux-4.7-rc3.orig/block/blk-sysfs.c 2016-06-13 17:19:43.000000000 +0200 +++ linux-4.7-rc3/block/blk-sysfs.c 2016-06-13 17:20:24.000000000 +0200 @@ -198,6 +198,13 @@ static ssize_t queue_copy_max_show(struc (unsigned long long)q->limits.max_copy_sectors << 9); } +static ssize_t queue_copy_boundary_show(struct request_queue *q, char *page) +{ + return sprintf(page, "%llu\n", + !q->limits.max_copy_sectors || q->limits.copy_boundary == 63 ? + 0ULL : 512ULL << q->limits.copy_boundary); +} + static ssize_t queue_max_sectors_store(struct request_queue *q, const char *page, size_t count) { @@ -484,6 +491,11 @@ static struct queue_sysfs_entry queue_co .show = queue_copy_max_show, }; +static struct queue_sysfs_entry queue_copy_boundary_entry = { + .attr = {.name = "copy_boundary_bytes", .mode = S_IRUGO }, + .show = queue_copy_boundary_show, +}; + static struct queue_sysfs_entry queue_nonrot_entry = { .attr = {.name = "rotational", .mode = S_IRUGO | S_IWUSR }, .show = queue_show_nonrot, @@ -546,6 +558,7 @@ static struct attribute *default_attrs[] &queue_discard_zeroes_data_entry.attr, &queue_write_same_max_entry.attr, &queue_copy_max_entry.attr, + &queue_copy_boundary_entry.attr, &queue_nonrot_entry.attr, &queue_nomerges_entry.attr, &queue_rq_affinity_entry.attr, Index: linux-4.7-rc3/include/linux/blkdev.h =================================================================== --- linux-4.7-rc3.orig/include/linux/blkdev.h 2016-06-13 17:19:44.000000000 +0200 +++ linux-4.7-rc3/include/linux/blkdev.h 2016-06-13 17:20:24.000000000 +0200 @@ -274,6 +274,7 @@ struct queue_limits { unsigned short max_segments; unsigned short max_integrity_segments; + unsigned char copy_boundary; unsigned char misaligned; unsigned char discard_misaligned; unsigned char cluster; Index: linux-4.7-rc3/block/blk-lib.c =================================================================== --- linux-4.7-rc3.orig/block/blk-lib.c 2016-06-13 17:19:44.000000000 +0200 +++ linux-4.7-rc3/block/blk-lib.c 2016-06-13 17:19:45.000000000 +0200 @@ -403,6 +403,12 @@ int blkdev_issue_copy(struct block_devic goto end_callback; } + /* + * TODO: if copy_boundary is too small, we'd better reject this + * request and let the caller perform copy manually. The threshold + * should be determined experimentally. + */ + if (src_sector + nr_sects < src_sector || dst_sector + nr_sects < dst_sector) { ret = -EINVAL; @@ -440,7 +446,20 @@ int blkdev_issue_copy(struct block_devic while (nr_sects && !ACCESS_ONCE(batch->async_error)) { struct bio *read_bio, *write_bio; struct bio_copy *bc; - unsigned chunk = (unsigned)min(nr_sects, (sector_t)max_copy_sectors); + unsigned chunk; + u64 boundary, max_io; + + chunk = (unsigned)min(nr_sects, (sector_t)max_copy_sectors); + + boundary = 1ULL << sq->limits.copy_boundary; + max_io = boundary - (src_sector & (boundary - 1)); + if (chunk > max_io) + chunk = max_io; + + boundary = 1ULL << dq->limits.copy_boundary; + max_io = boundary - (dst_sector & (boundary - 1)); + if (chunk > max_io) + chunk = max_io; bc = kmalloc(sizeof(struct bio_copy), gfp_mask); if (!bc) {