Commit 49077326 authored by NeilBrown's avatar NeilBrown

md: move io accounting out of personalities into md_make_request

While I generally prefer letting personalities do as much as possible,
given that we have a central md_make_request anyway we may as well use
it to simplify code.
Also this centralises knowledge of ->gendisk which will help later.
Signed-off-by: default avatarNeilBrown <neilb@suse.de>
parent 2b7f2228
...@@ -288,23 +288,15 @@ static int linear_stop (mddev_t *mddev) ...@@ -288,23 +288,15 @@ static int linear_stop (mddev_t *mddev)
static int linear_make_request (struct request_queue *q, struct bio *bio) static int linear_make_request (struct request_queue *q, struct bio *bio)
{ {
const int rw = bio_data_dir(bio);
mddev_t *mddev = q->queuedata; mddev_t *mddev = q->queuedata;
dev_info_t *tmp_dev; dev_info_t *tmp_dev;
sector_t start_sector; sector_t start_sector;
int cpu;
if (unlikely(bio_rw_flagged(bio, BIO_RW_BARRIER))) { if (unlikely(bio_rw_flagged(bio, BIO_RW_BARRIER))) {
md_barrier_request(mddev, bio); md_barrier_request(mddev, bio);
return 0; return 0;
} }
cpu = part_stat_lock();
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
part_stat_add(cpu, &mddev->gendisk->part0, sectors[rw],
bio_sectors(bio));
part_stat_unlock();
rcu_read_lock(); rcu_read_lock();
tmp_dev = which_dev(mddev, bio->bi_sector); tmp_dev = which_dev(mddev, bio->bi_sector);
start_sector = tmp_dev->end_sector - tmp_dev->rdev->sectors; start_sector = tmp_dev->end_sector - tmp_dev->rdev->sectors;
......
...@@ -214,8 +214,11 @@ static DEFINE_SPINLOCK(all_mddevs_lock); ...@@ -214,8 +214,11 @@ static DEFINE_SPINLOCK(all_mddevs_lock);
*/ */
static int md_make_request(struct request_queue *q, struct bio *bio) static int md_make_request(struct request_queue *q, struct bio *bio)
{ {
const int rw = bio_data_dir(bio);
mddev_t *mddev = q->queuedata; mddev_t *mddev = q->queuedata;
int rv; int rv;
int cpu;
if (mddev == NULL || mddev->pers == NULL) { if (mddev == NULL || mddev->pers == NULL) {
bio_io_error(bio); bio_io_error(bio);
return 0; return 0;
...@@ -236,7 +239,15 @@ static int md_make_request(struct request_queue *q, struct bio *bio) ...@@ -236,7 +239,15 @@ static int md_make_request(struct request_queue *q, struct bio *bio)
} }
atomic_inc(&mddev->active_io); atomic_inc(&mddev->active_io);
rcu_read_unlock(); rcu_read_unlock();
rv = mddev->pers->make_request(q, bio); rv = mddev->pers->make_request(q, bio);
cpu = part_stat_lock();
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
part_stat_add(cpu, &mddev->gendisk->part0, sectors[rw],
bio_sectors(bio));
part_stat_unlock();
if (atomic_dec_and_test(&mddev->active_io) && mddev->suspended) if (atomic_dec_and_test(&mddev->active_io) && mddev->suspended)
wake_up(&mddev->sb_wait); wake_up(&mddev->sb_wait);
......
...@@ -141,8 +141,6 @@ static int multipath_make_request (struct request_queue *q, struct bio * bio) ...@@ -141,8 +141,6 @@ static int multipath_make_request (struct request_queue *q, struct bio * bio)
multipath_conf_t *conf = mddev->private; multipath_conf_t *conf = mddev->private;
struct multipath_bh * mp_bh; struct multipath_bh * mp_bh;
struct multipath_info *multipath; struct multipath_info *multipath;
const int rw = bio_data_dir(bio);
int cpu;
if (unlikely(bio_rw_flagged(bio, BIO_RW_BARRIER))) { if (unlikely(bio_rw_flagged(bio, BIO_RW_BARRIER))) {
md_barrier_request(mddev, bio); md_barrier_request(mddev, bio);
...@@ -154,12 +152,6 @@ static int multipath_make_request (struct request_queue *q, struct bio * bio) ...@@ -154,12 +152,6 @@ static int multipath_make_request (struct request_queue *q, struct bio * bio)
mp_bh->master_bio = bio; mp_bh->master_bio = bio;
mp_bh->mddev = mddev; mp_bh->mddev = mddev;
cpu = part_stat_lock();
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
part_stat_add(cpu, &mddev->gendisk->part0, sectors[rw],
bio_sectors(bio));
part_stat_unlock();
mp_bh->path = multipath_map(conf); mp_bh->path = multipath_map(conf);
if (mp_bh->path < 0) { if (mp_bh->path < 0) {
bio_endio(bio, -EIO); bio_endio(bio, -EIO);
......
...@@ -472,20 +472,12 @@ static int raid0_make_request(struct request_queue *q, struct bio *bio) ...@@ -472,20 +472,12 @@ static int raid0_make_request(struct request_queue *q, struct bio *bio)
sector_t sector_offset; sector_t sector_offset;
struct strip_zone *zone; struct strip_zone *zone;
mdk_rdev_t *tmp_dev; mdk_rdev_t *tmp_dev;
const int rw = bio_data_dir(bio);
int cpu;
if (unlikely(bio_rw_flagged(bio, BIO_RW_BARRIER))) { if (unlikely(bio_rw_flagged(bio, BIO_RW_BARRIER))) {
md_barrier_request(mddev, bio); md_barrier_request(mddev, bio);
return 0; return 0;
} }
cpu = part_stat_lock();
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
part_stat_add(cpu, &mddev->gendisk->part0, sectors[rw],
bio_sectors(bio));
part_stat_unlock();
chunk_sects = mddev->chunk_sectors; chunk_sects = mddev->chunk_sectors;
if (unlikely(!is_io_in_chunk_boundary(mddev, chunk_sects, bio))) { if (unlikely(!is_io_in_chunk_boundary(mddev, chunk_sects, bio))) {
sector_t sector = bio->bi_sector; sector_t sector = bio->bi_sector;
......
...@@ -787,7 +787,6 @@ static int make_request(struct request_queue *q, struct bio * bio) ...@@ -787,7 +787,6 @@ static int make_request(struct request_queue *q, struct bio * bio)
struct page **behind_pages = NULL; struct page **behind_pages = NULL;
const int rw = bio_data_dir(bio); const int rw = bio_data_dir(bio);
const bool do_sync = bio_rw_flagged(bio, BIO_RW_SYNCIO); const bool do_sync = bio_rw_flagged(bio, BIO_RW_SYNCIO);
int cpu;
bool do_barriers; bool do_barriers;
mdk_rdev_t *blocked_rdev; mdk_rdev_t *blocked_rdev;
...@@ -833,12 +832,6 @@ static int make_request(struct request_queue *q, struct bio * bio) ...@@ -833,12 +832,6 @@ static int make_request(struct request_queue *q, struct bio * bio)
bitmap = mddev->bitmap; bitmap = mddev->bitmap;
cpu = part_stat_lock();
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
part_stat_add(cpu, &mddev->gendisk->part0, sectors[rw],
bio_sectors(bio));
part_stat_unlock();
/* /*
* make_request() can abort the operation when READA is being * make_request() can abort the operation when READA is being
* used and no empty request is available. * used and no empty request is available.
......
...@@ -795,7 +795,6 @@ static int make_request(struct request_queue *q, struct bio * bio) ...@@ -795,7 +795,6 @@ static int make_request(struct request_queue *q, struct bio * bio)
mirror_info_t *mirror; mirror_info_t *mirror;
r10bio_t *r10_bio; r10bio_t *r10_bio;
struct bio *read_bio; struct bio *read_bio;
int cpu;
int i; int i;
int chunk_sects = conf->chunk_mask + 1; int chunk_sects = conf->chunk_mask + 1;
const int rw = bio_data_dir(bio); const int rw = bio_data_dir(bio);
...@@ -850,12 +849,6 @@ static int make_request(struct request_queue *q, struct bio * bio) ...@@ -850,12 +849,6 @@ static int make_request(struct request_queue *q, struct bio * bio)
*/ */
wait_barrier(conf); wait_barrier(conf);
cpu = part_stat_lock();
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
part_stat_add(cpu, &mddev->gendisk->part0, sectors[rw],
bio_sectors(bio));
part_stat_unlock();
r10_bio = mempool_alloc(conf->r10bio_pool, GFP_NOIO); r10_bio = mempool_alloc(conf->r10bio_pool, GFP_NOIO);
r10_bio->master_bio = bio; r10_bio->master_bio = bio;
......
...@@ -3879,7 +3879,7 @@ static int make_request(struct request_queue *q, struct bio * bi) ...@@ -3879,7 +3879,7 @@ static int make_request(struct request_queue *q, struct bio * bi)
sector_t logical_sector, last_sector; sector_t logical_sector, last_sector;
struct stripe_head *sh; struct stripe_head *sh;
const int rw = bio_data_dir(bi); const int rw = bio_data_dir(bi);
int cpu, remaining; int remaining;
if (unlikely(bio_rw_flagged(bi, BIO_RW_BARRIER))) { if (unlikely(bio_rw_flagged(bi, BIO_RW_BARRIER))) {
/* Drain all pending writes. We only really need /* Drain all pending writes. We only really need
...@@ -3894,12 +3894,6 @@ static int make_request(struct request_queue *q, struct bio * bi) ...@@ -3894,12 +3894,6 @@ static int make_request(struct request_queue *q, struct bio * bi)
md_write_start(mddev, bi); md_write_start(mddev, bi);
cpu = part_stat_lock();
part_stat_inc(cpu, &mddev->gendisk->part0, ios[rw]);
part_stat_add(cpu, &mddev->gendisk->part0, sectors[rw],
bio_sectors(bi));
part_stat_unlock();
if (rw == READ && if (rw == READ &&
mddev->reshape_position == MaxSector && mddev->reshape_position == MaxSector &&
chunk_aligned_read(q,bi)) chunk_aligned_read(q,bi))
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment