dm: Refactor for new bio cloning/splitting
We need to convert the dm code to the new bvec_iter primitives which respect bi_bvec_done; they also allow us to drastically simplify dm's bio splitting code. Also, it's no longer necessary to save/restore the bvec array anymore - driver conversions for immutable bvecs are done, so drivers should never be modifying it. Also kill bio_sector_offset(), dm was the only user and it doesn't make much sense anymore. Signed-off-by: Kent Overstreet <kmo@daterainc.com> Cc: Jens Axboe <axboe@kernel.dk> Cc: Alasdair Kergon <agk@redhat.com> Cc: dm-devel@redhat.com Reviewed-by: Mike Snitzer <snitzer@redhat.com>
This commit is contained in:
parent
5341a6278b
commit
1c3b13e64c
@ -17,49 +17,24 @@
|
|||||||
* original bio state.
|
* original bio state.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
struct dm_bio_vec_details {
|
|
||||||
#if PAGE_SIZE < 65536
|
|
||||||
__u16 bv_len;
|
|
||||||
__u16 bv_offset;
|
|
||||||
#else
|
|
||||||
unsigned bv_len;
|
|
||||||
unsigned bv_offset;
|
|
||||||
#endif
|
|
||||||
};
|
|
||||||
|
|
||||||
struct dm_bio_details {
|
struct dm_bio_details {
|
||||||
struct block_device *bi_bdev;
|
struct block_device *bi_bdev;
|
||||||
unsigned long bi_flags;
|
unsigned long bi_flags;
|
||||||
struct bvec_iter bi_iter;
|
struct bvec_iter bi_iter;
|
||||||
struct dm_bio_vec_details bi_io_vec[BIO_MAX_PAGES];
|
|
||||||
};
|
};
|
||||||
|
|
||||||
static inline void dm_bio_record(struct dm_bio_details *bd, struct bio *bio)
|
static inline void dm_bio_record(struct dm_bio_details *bd, struct bio *bio)
|
||||||
{
|
{
|
||||||
unsigned i;
|
|
||||||
|
|
||||||
bd->bi_bdev = bio->bi_bdev;
|
bd->bi_bdev = bio->bi_bdev;
|
||||||
bd->bi_flags = bio->bi_flags;
|
bd->bi_flags = bio->bi_flags;
|
||||||
bd->bi_iter = bio->bi_iter;
|
bd->bi_iter = bio->bi_iter;
|
||||||
|
|
||||||
for (i = 0; i < bio->bi_vcnt; i++) {
|
|
||||||
bd->bi_io_vec[i].bv_len = bio->bi_io_vec[i].bv_len;
|
|
||||||
bd->bi_io_vec[i].bv_offset = bio->bi_io_vec[i].bv_offset;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void dm_bio_restore(struct dm_bio_details *bd, struct bio *bio)
|
static inline void dm_bio_restore(struct dm_bio_details *bd, struct bio *bio)
|
||||||
{
|
{
|
||||||
unsigned i;
|
|
||||||
|
|
||||||
bio->bi_bdev = bd->bi_bdev;
|
bio->bi_bdev = bd->bi_bdev;
|
||||||
bio->bi_flags = bd->bi_flags;
|
bio->bi_flags = bd->bi_flags;
|
||||||
bio->bi_iter = bd->bi_iter;
|
bio->bi_iter = bd->bi_iter;
|
||||||
|
|
||||||
for (i = 0; i < bio->bi_vcnt; i++) {
|
|
||||||
bio->bi_io_vec[i].bv_len = bd->bi_io_vec[i].bv_len;
|
|
||||||
bio->bi_io_vec[i].bv_offset = bd->bi_io_vec[i].bv_offset;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
174
drivers/md/dm.c
174
drivers/md/dm.c
@ -1155,7 +1155,6 @@ struct clone_info {
|
|||||||
struct dm_io *io;
|
struct dm_io *io;
|
||||||
sector_t sector;
|
sector_t sector;
|
||||||
sector_t sector_count;
|
sector_t sector_count;
|
||||||
unsigned short idx;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
static void bio_setup_sector(struct bio *bio, sector_t sector, sector_t len)
|
static void bio_setup_sector(struct bio *bio, sector_t sector, sector_t len)
|
||||||
@ -1164,68 +1163,24 @@ static void bio_setup_sector(struct bio *bio, sector_t sector, sector_t len)
|
|||||||
bio->bi_iter.bi_size = to_bytes(len);
|
bio->bi_iter.bi_size = to_bytes(len);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void bio_setup_bv(struct bio *bio, unsigned short idx, unsigned short bv_count)
|
|
||||||
{
|
|
||||||
bio->bi_iter.bi_idx = idx;
|
|
||||||
bio->bi_vcnt = idx + bv_count;
|
|
||||||
bio->bi_flags &= ~(1 << BIO_SEG_VALID);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void clone_bio_integrity(struct bio *bio, struct bio *clone,
|
|
||||||
unsigned short idx, unsigned len, unsigned offset,
|
|
||||||
unsigned trim)
|
|
||||||
{
|
|
||||||
if (!bio_integrity(bio))
|
|
||||||
return;
|
|
||||||
|
|
||||||
bio_integrity_clone(clone, bio, GFP_NOIO);
|
|
||||||
|
|
||||||
if (trim)
|
|
||||||
bio_integrity_trim(clone, bio_sector_offset(bio, idx, offset), len);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Creates a little bio that just does part of a bvec.
|
|
||||||
*/
|
|
||||||
static void clone_split_bio(struct dm_target_io *tio, struct bio *bio,
|
|
||||||
sector_t sector, unsigned short idx,
|
|
||||||
unsigned offset, unsigned len)
|
|
||||||
{
|
|
||||||
struct bio *clone = &tio->clone;
|
|
||||||
struct bio_vec *bv = bio->bi_io_vec + idx;
|
|
||||||
|
|
||||||
*clone->bi_io_vec = *bv;
|
|
||||||
|
|
||||||
bio_setup_sector(clone, sector, len);
|
|
||||||
|
|
||||||
clone->bi_bdev = bio->bi_bdev;
|
|
||||||
clone->bi_rw = bio->bi_rw;
|
|
||||||
clone->bi_vcnt = 1;
|
|
||||||
clone->bi_io_vec->bv_offset = offset;
|
|
||||||
clone->bi_io_vec->bv_len = clone->bi_iter.bi_size;
|
|
||||||
clone->bi_flags |= 1 << BIO_CLONED;
|
|
||||||
|
|
||||||
clone_bio_integrity(bio, clone, idx, len, offset, 1);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Creates a bio that consists of range of complete bvecs.
|
* Creates a bio that consists of range of complete bvecs.
|
||||||
*/
|
*/
|
||||||
static void clone_bio(struct dm_target_io *tio, struct bio *bio,
|
static void clone_bio(struct dm_target_io *tio, struct bio *bio,
|
||||||
sector_t sector, unsigned short idx,
|
sector_t sector, unsigned len)
|
||||||
unsigned short bv_count, unsigned len)
|
|
||||||
{
|
{
|
||||||
struct bio *clone = &tio->clone;
|
struct bio *clone = &tio->clone;
|
||||||
unsigned trim = 0;
|
|
||||||
|
|
||||||
__bio_clone(clone, bio);
|
__bio_clone_fast(clone, bio);
|
||||||
bio_setup_sector(clone, sector, len);
|
|
||||||
bio_setup_bv(clone, idx, bv_count);
|
|
||||||
|
|
||||||
if (idx != bio->bi_iter.bi_idx ||
|
if (bio_integrity(bio))
|
||||||
clone->bi_iter.bi_size < bio->bi_iter.bi_size)
|
bio_integrity_clone(clone, bio, GFP_NOIO);
|
||||||
trim = 1;
|
|
||||||
clone_bio_integrity(bio, clone, idx, len, 0, trim);
|
bio_advance(clone, to_bytes(sector - clone->bi_iter.bi_sector));
|
||||||
|
clone->bi_iter.bi_size = to_bytes(len);
|
||||||
|
|
||||||
|
if (bio_integrity(bio))
|
||||||
|
bio_integrity_trim(clone, 0, len);
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct dm_target_io *alloc_tio(struct clone_info *ci,
|
static struct dm_target_io *alloc_tio(struct clone_info *ci,
|
||||||
@ -1258,7 +1213,7 @@ static void __clone_and_map_simple_bio(struct clone_info *ci,
|
|||||||
* ci->bio->bi_max_vecs is BIO_INLINE_VECS anyway, for both flush
|
* ci->bio->bi_max_vecs is BIO_INLINE_VECS anyway, for both flush
|
||||||
* and discard, so no need for concern about wasted bvec allocations.
|
* and discard, so no need for concern about wasted bvec allocations.
|
||||||
*/
|
*/
|
||||||
__bio_clone(clone, ci->bio);
|
__bio_clone_fast(clone, ci->bio);
|
||||||
if (len)
|
if (len)
|
||||||
bio_setup_sector(clone, ci->sector, len);
|
bio_setup_sector(clone, ci->sector, len);
|
||||||
|
|
||||||
@ -1287,10 +1242,7 @@ static int __send_empty_flush(struct clone_info *ci)
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void __clone_and_map_data_bio(struct clone_info *ci, struct dm_target *ti,
|
static void __clone_and_map_data_bio(struct clone_info *ci, struct dm_target *ti,
|
||||||
sector_t sector, int nr_iovecs,
|
sector_t sector, unsigned len)
|
||||||
unsigned short idx, unsigned short bv_count,
|
|
||||||
unsigned offset, unsigned len,
|
|
||||||
unsigned split_bvec)
|
|
||||||
{
|
{
|
||||||
struct bio *bio = ci->bio;
|
struct bio *bio = ci->bio;
|
||||||
struct dm_target_io *tio;
|
struct dm_target_io *tio;
|
||||||
@ -1304,11 +1256,8 @@ static void __clone_and_map_data_bio(struct clone_info *ci, struct dm_target *ti
|
|||||||
num_target_bios = ti->num_write_bios(ti, bio);
|
num_target_bios = ti->num_write_bios(ti, bio);
|
||||||
|
|
||||||
for (target_bio_nr = 0; target_bio_nr < num_target_bios; target_bio_nr++) {
|
for (target_bio_nr = 0; target_bio_nr < num_target_bios; target_bio_nr++) {
|
||||||
tio = alloc_tio(ci, ti, nr_iovecs, target_bio_nr);
|
tio = alloc_tio(ci, ti, 0, target_bio_nr);
|
||||||
if (split_bvec)
|
clone_bio(tio, bio, sector, len);
|
||||||
clone_split_bio(tio, bio, sector, idx, offset, len);
|
|
||||||
else
|
|
||||||
clone_bio(tio, bio, sector, idx, bv_count, len);
|
|
||||||
__map_bio(tio);
|
__map_bio(tio);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1379,60 +1328,6 @@ static int __send_write_same(struct clone_info *ci)
|
|||||||
return __send_changing_extent_only(ci, get_num_write_same_bios, NULL);
|
return __send_changing_extent_only(ci, get_num_write_same_bios, NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* Find maximum number of sectors / bvecs we can process with a single bio.
|
|
||||||
*/
|
|
||||||
static sector_t __len_within_target(struct clone_info *ci, sector_t max, int *idx)
|
|
||||||
{
|
|
||||||
struct bio *bio = ci->bio;
|
|
||||||
sector_t bv_len, total_len = 0;
|
|
||||||
|
|
||||||
for (*idx = ci->idx; max && (*idx < bio->bi_vcnt); (*idx)++) {
|
|
||||||
bv_len = to_sector(bio->bi_io_vec[*idx].bv_len);
|
|
||||||
|
|
||||||
if (bv_len > max)
|
|
||||||
break;
|
|
||||||
|
|
||||||
max -= bv_len;
|
|
||||||
total_len += bv_len;
|
|
||||||
}
|
|
||||||
|
|
||||||
return total_len;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int __split_bvec_across_targets(struct clone_info *ci,
|
|
||||||
struct dm_target *ti, sector_t max)
|
|
||||||
{
|
|
||||||
struct bio *bio = ci->bio;
|
|
||||||
struct bio_vec *bv = bio->bi_io_vec + ci->idx;
|
|
||||||
sector_t remaining = to_sector(bv->bv_len);
|
|
||||||
unsigned offset = 0;
|
|
||||||
sector_t len;
|
|
||||||
|
|
||||||
do {
|
|
||||||
if (offset) {
|
|
||||||
ti = dm_table_find_target(ci->map, ci->sector);
|
|
||||||
if (!dm_target_is_valid(ti))
|
|
||||||
return -EIO;
|
|
||||||
|
|
||||||
max = max_io_len(ci->sector, ti);
|
|
||||||
}
|
|
||||||
|
|
||||||
len = min(remaining, max);
|
|
||||||
|
|
||||||
__clone_and_map_data_bio(ci, ti, ci->sector, 1, ci->idx, 0,
|
|
||||||
bv->bv_offset + offset, len, 1);
|
|
||||||
|
|
||||||
ci->sector += len;
|
|
||||||
ci->sector_count -= len;
|
|
||||||
offset += to_bytes(len);
|
|
||||||
} while (remaining -= len);
|
|
||||||
|
|
||||||
ci->idx++;
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Select the correct strategy for processing a non-flush bio.
|
* Select the correct strategy for processing a non-flush bio.
|
||||||
*/
|
*/
|
||||||
@ -1440,8 +1335,7 @@ static int __split_and_process_non_flush(struct clone_info *ci)
|
|||||||
{
|
{
|
||||||
struct bio *bio = ci->bio;
|
struct bio *bio = ci->bio;
|
||||||
struct dm_target *ti;
|
struct dm_target *ti;
|
||||||
sector_t len, max;
|
unsigned len;
|
||||||
int idx;
|
|
||||||
|
|
||||||
if (unlikely(bio->bi_rw & REQ_DISCARD))
|
if (unlikely(bio->bi_rw & REQ_DISCARD))
|
||||||
return __send_discard(ci);
|
return __send_discard(ci);
|
||||||
@ -1452,41 +1346,14 @@ static int __split_and_process_non_flush(struct clone_info *ci)
|
|||||||
if (!dm_target_is_valid(ti))
|
if (!dm_target_is_valid(ti))
|
||||||
return -EIO;
|
return -EIO;
|
||||||
|
|
||||||
max = max_io_len(ci->sector, ti);
|
len = min_t(sector_t, max_io_len(ci->sector, ti), ci->sector_count);
|
||||||
|
|
||||||
/*
|
__clone_and_map_data_bio(ci, ti, ci->sector, len);
|
||||||
* Optimise for the simple case where we can do all of
|
|
||||||
* the remaining io with a single clone.
|
|
||||||
*/
|
|
||||||
if (ci->sector_count <= max) {
|
|
||||||
__clone_and_map_data_bio(ci, ti, ci->sector, bio->bi_max_vecs,
|
|
||||||
ci->idx, bio->bi_vcnt - ci->idx, 0,
|
|
||||||
ci->sector_count, 0);
|
|
||||||
ci->sector_count = 0;
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
ci->sector += len;
|
||||||
* There are some bvecs that don't span targets.
|
ci->sector_count -= len;
|
||||||
* Do as many of these as possible.
|
|
||||||
*/
|
|
||||||
if (to_sector(bio->bi_io_vec[ci->idx].bv_len) <= max) {
|
|
||||||
len = __len_within_target(ci, max, &idx);
|
|
||||||
|
|
||||||
__clone_and_map_data_bio(ci, ti, ci->sector, bio->bi_max_vecs,
|
return 0;
|
||||||
ci->idx, idx - ci->idx, 0, len, 0);
|
|
||||||
|
|
||||||
ci->sector += len;
|
|
||||||
ci->sector_count -= len;
|
|
||||||
ci->idx = idx;
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Handle a bvec that must be split between two or more targets.
|
|
||||||
*/
|
|
||||||
return __split_bvec_across_targets(ci, ti, max);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -1512,7 +1379,6 @@ static void __split_and_process_bio(struct mapped_device *md,
|
|||||||
ci.io->md = md;
|
ci.io->md = md;
|
||||||
spin_lock_init(&ci.io->endio_lock);
|
spin_lock_init(&ci.io->endio_lock);
|
||||||
ci.sector = bio->bi_iter.bi_sector;
|
ci.sector = bio->bi_iter.bi_sector;
|
||||||
ci.idx = bio->bi_iter.bi_idx;
|
|
||||||
|
|
||||||
start_io_acct(ci.io);
|
start_io_acct(ci.io);
|
||||||
|
|
||||||
|
72
fs/bio.c
72
fs/bio.c
@ -514,40 +514,6 @@ inline int bio_phys_segments(struct request_queue *q, struct bio *bio)
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL(bio_phys_segments);
|
EXPORT_SYMBOL(bio_phys_segments);
|
||||||
|
|
||||||
/**
|
|
||||||
* __bio_clone - clone a bio
|
|
||||||
* @bio: destination bio
|
|
||||||
* @bio_src: bio to clone
|
|
||||||
*
|
|
||||||
* Clone a &bio. Caller will own the returned bio, but not
|
|
||||||
* the actual data it points to. Reference count of returned
|
|
||||||
* bio will be one.
|
|
||||||
*/
|
|
||||||
void __bio_clone(struct bio *bio, struct bio *bio_src)
|
|
||||||
{
|
|
||||||
if (bio_is_rw(bio_src)) {
|
|
||||||
struct bio_vec bv;
|
|
||||||
struct bvec_iter iter;
|
|
||||||
|
|
||||||
bio_for_each_segment(bv, bio_src, iter)
|
|
||||||
bio->bi_io_vec[bio->bi_vcnt++] = bv;
|
|
||||||
} else if (bio_has_data(bio_src)) {
|
|
||||||
memcpy(bio->bi_io_vec, bio_src->bi_io_vec,
|
|
||||||
bio_src->bi_max_vecs * sizeof(struct bio_vec));
|
|
||||||
bio->bi_vcnt = bio_src->bi_vcnt;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* most users will be overriding ->bi_bdev with a new target,
|
|
||||||
* so we don't set nor calculate new physical/hw segment counts here
|
|
||||||
*/
|
|
||||||
bio->bi_bdev = bio_src->bi_bdev;
|
|
||||||
bio->bi_flags |= 1 << BIO_CLONED;
|
|
||||||
bio->bi_rw = bio_src->bi_rw;
|
|
||||||
bio->bi_iter = bio_src->bi_iter;
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(__bio_clone);
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* __bio_clone_fast - clone a bio that shares the original bio's biovec
|
* __bio_clone_fast - clone a bio that shares the original bio's biovec
|
||||||
* @bio: destination bio
|
* @bio: destination bio
|
||||||
@ -1921,44 +1887,6 @@ void bio_trim(struct bio *bio, int offset, int size)
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(bio_trim);
|
EXPORT_SYMBOL_GPL(bio_trim);
|
||||||
|
|
||||||
/**
|
|
||||||
* bio_sector_offset - Find hardware sector offset in bio
|
|
||||||
* @bio: bio to inspect
|
|
||||||
* @index: bio_vec index
|
|
||||||
* @offset: offset in bv_page
|
|
||||||
*
|
|
||||||
* Return the number of hardware sectors between beginning of bio
|
|
||||||
* and an end point indicated by a bio_vec index and an offset
|
|
||||||
* within that vector's page.
|
|
||||||
*/
|
|
||||||
sector_t bio_sector_offset(struct bio *bio, unsigned short index,
|
|
||||||
unsigned int offset)
|
|
||||||
{
|
|
||||||
unsigned int sector_sz;
|
|
||||||
struct bio_vec *bv;
|
|
||||||
sector_t sectors;
|
|
||||||
int i;
|
|
||||||
|
|
||||||
sector_sz = queue_logical_block_size(bio->bi_bdev->bd_disk->queue);
|
|
||||||
sectors = 0;
|
|
||||||
|
|
||||||
if (index >= bio->bi_iter.bi_idx)
|
|
||||||
index = bio->bi_vcnt - 1;
|
|
||||||
|
|
||||||
bio_for_each_segment_all(bv, bio, i) {
|
|
||||||
if (i == index) {
|
|
||||||
if (offset > bv->bv_offset)
|
|
||||||
sectors += (offset - bv->bv_offset) / sector_sz;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
sectors += bv->bv_len / sector_sz;
|
|
||||||
}
|
|
||||||
|
|
||||||
return sectors;
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(bio_sector_offset);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* create memory pools for biovec's in a bio_set.
|
* create memory pools for biovec's in a bio_set.
|
||||||
* use the global biovec slabs created for general use.
|
* use the global biovec slabs created for general use.
|
||||||
|
@ -330,7 +330,6 @@ extern void bio_put(struct bio *);
|
|||||||
|
|
||||||
extern void __bio_clone_fast(struct bio *, struct bio *);
|
extern void __bio_clone_fast(struct bio *, struct bio *);
|
||||||
extern struct bio *bio_clone_fast(struct bio *, gfp_t, struct bio_set *);
|
extern struct bio *bio_clone_fast(struct bio *, gfp_t, struct bio_set *);
|
||||||
extern void __bio_clone(struct bio *, struct bio *);
|
|
||||||
extern struct bio *bio_clone_bioset(struct bio *, gfp_t, struct bio_set *bs);
|
extern struct bio *bio_clone_bioset(struct bio *, gfp_t, struct bio_set *bs);
|
||||||
|
|
||||||
extern struct bio_set *fs_bio_set;
|
extern struct bio_set *fs_bio_set;
|
||||||
@ -370,7 +369,6 @@ extern int bio_add_page(struct bio *, struct page *, unsigned int,unsigned int);
|
|||||||
extern int bio_add_pc_page(struct request_queue *, struct bio *, struct page *,
|
extern int bio_add_pc_page(struct request_queue *, struct bio *, struct page *,
|
||||||
unsigned int, unsigned int);
|
unsigned int, unsigned int);
|
||||||
extern int bio_get_nr_vecs(struct block_device *);
|
extern int bio_get_nr_vecs(struct block_device *);
|
||||||
extern sector_t bio_sector_offset(struct bio *, unsigned short, unsigned int);
|
|
||||||
extern struct bio *bio_map_user(struct request_queue *, struct block_device *,
|
extern struct bio *bio_map_user(struct request_queue *, struct block_device *,
|
||||||
unsigned long, unsigned int, int, gfp_t);
|
unsigned long, unsigned int, int, gfp_t);
|
||||||
struct sg_iovec;
|
struct sg_iovec;
|
||||||
|
Loading…
Reference in New Issue
Block a user