dm: move bio_io_error into __split_and_process_bio

Move the bio_io_error() calls directly into __split_and_process_bio().

This avoids some code duplication in later patches.

Signed-off-by: Mikulas Patocka <mpatocka@redhat.com>
Signed-off-by: Alasdair G Kergon <agk@redhat.com>
This commit is contained in:
Mikulas Patocka 2009-04-02 19:55:38 +01:00 committed by Alasdair G Kergon
parent 8a53c28db4
commit f0b9a4502b
1 changed files with 10 additions and 11 deletions

View File

@ -828,18 +828,20 @@ static int __clone_and_map(struct clone_info *ci)
/*
* Split the bio into several clones and submit it to targets.
*/
static int __split_and_process_bio(struct mapped_device *md, struct bio *bio)
static void __split_and_process_bio(struct mapped_device *md, struct bio *bio)
{
struct clone_info ci;
int error = 0;
ci.map = dm_get_table(md);
if (unlikely(!ci.map))
return -EIO;
if (unlikely(!ci.map)) {
bio_io_error(bio);
return;
}
if (unlikely(bio_barrier(bio) && !dm_table_barrier_ok(ci.map))) {
dm_table_put(ci.map);
bio_endio(bio, -EOPNOTSUPP);
return 0;
return;
}
ci.md = md;
ci.bio = bio;
@ -859,8 +861,6 @@ static int __split_and_process_bio(struct mapped_device *md, struct bio *bio)
/* drop the extra reference count */
dec_pending(ci.io, error);
dm_table_put(ci.map);
return 0;
}
/*-----------------------------------------------------------------
* CRUD END
@ -951,8 +951,9 @@ static int dm_request(struct request_queue *q, struct bio *bio)
down_read(&md->io_lock);
}
r = __split_and_process_bio(md, bio);
__split_and_process_bio(md, bio);
up_read(&md->io_lock);
return 0;
out_req:
if (r < 0)
@ -1404,10 +1405,8 @@ static void __flush_deferred_io(struct mapped_device *md)
{
struct bio *c;
while ((c = bio_list_pop(&md->deferred))) {
if (__split_and_process_bio(md, c))
bio_io_error(c);
}
while ((c = bio_list_pop(&md->deferred)))
__split_and_process_bio(md, c);
clear_bit(DMF_BLOCK_IO, &md->flags);
}