Loading drivers/md/raid0.c +18 −1 Original line number Original line Diff line number Diff line Loading @@ -88,6 +88,7 @@ static int create_strip_zones(struct mddev *mddev, struct r0conf **private_conf) char b[BDEVNAME_SIZE]; char b[BDEVNAME_SIZE]; char b2[BDEVNAME_SIZE]; char b2[BDEVNAME_SIZE]; struct r0conf *conf = kzalloc(sizeof(*conf), GFP_KERNEL); struct r0conf *conf = kzalloc(sizeof(*conf), GFP_KERNEL); bool discard_supported = false; if (!conf) if (!conf) return -ENOMEM; return -ENOMEM; Loading Loading @@ -195,6 +196,9 @@ static int create_strip_zones(struct mddev *mddev, struct r0conf **private_conf) if (!smallest || (rdev1->sectors < smallest->sectors)) if (!smallest || (rdev1->sectors < smallest->sectors)) smallest = rdev1; smallest = rdev1; cnt++; cnt++; if (blk_queue_discard(bdev_get_queue(rdev1->bdev))) discard_supported = true; } } if (cnt != mddev->raid_disks) { if (cnt != mddev->raid_disks) { printk(KERN_ERR "md/raid0:%s: too few disks (%d of %d) - " printk(KERN_ERR "md/raid0:%s: too few disks (%d of %d) - " Loading Loading @@ -272,6 +276,11 @@ static int create_strip_zones(struct mddev *mddev, struct r0conf **private_conf) blk_queue_io_opt(mddev->queue, blk_queue_io_opt(mddev->queue, (mddev->chunk_sectors << 9) * mddev->raid_disks); (mddev->chunk_sectors << 9) * mddev->raid_disks); if (!discard_supported) queue_flag_clear_unlocked(QUEUE_FLAG_DISCARD, mddev->queue); else queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, mddev->queue); pr_debug("md/raid0:%s: done.\n", mdname(mddev)); pr_debug("md/raid0:%s: done.\n", mdname(mddev)); *private_conf = conf; *private_conf = conf; Loading Loading @@ -423,6 +432,7 @@ static int raid0_run(struct mddev *mddev) return -EINVAL; return -EINVAL; blk_queue_max_hw_sectors(mddev->queue, mddev->chunk_sectors); blk_queue_max_hw_sectors(mddev->queue, mddev->chunk_sectors); blk_queue_max_write_same_sectors(mddev->queue, mddev->chunk_sectors); blk_queue_max_write_same_sectors(mddev->queue, mddev->chunk_sectors); blk_queue_max_discard_sectors(mddev->queue, mddev->chunk_sectors); /* if private is not null, we are here after takeover */ /* if private is not null, we are here after takeover */ if (mddev->private == NULL) { if (mddev->private == NULL) { Loading Loading @@ -510,7 +520,7 @@ static void raid0_make_request(struct mddev *mddev, struct bio *bio) sector_t sector = bio->bi_sector; sector_t sector = bio->bi_sector; struct bio_pair *bp; struct bio_pair *bp; /* Sanity check -- queue functions should prevent this happening */ /* Sanity check -- queue functions should prevent this happening */ if (bio->bi_vcnt != 1 || if ((bio->bi_vcnt != 1 && bio->bi_vcnt != 0) || bio->bi_idx != 0) bio->bi_idx != 0) goto bad_map; goto bad_map; /* This is a one page bio that upper layers /* This is a one page bio that upper layers Loading @@ -536,6 +546,13 @@ static void raid0_make_request(struct mddev *mddev, struct bio *bio) bio->bi_sector = sector_offset + zone->dev_start + bio->bi_sector = sector_offset + zone->dev_start + tmp_dev->data_offset; tmp_dev->data_offset; if (unlikely((bio->bi_rw & REQ_DISCARD) && !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) { /* Just ignore it */ bio_endio(bio, 0); return; } generic_make_request(bio); generic_make_request(bio); return; return; Loading Loading
drivers/md/raid0.c +18 −1 Original line number Original line Diff line number Diff line Loading @@ -88,6 +88,7 @@ static int create_strip_zones(struct mddev *mddev, struct r0conf **private_conf) char b[BDEVNAME_SIZE]; char b[BDEVNAME_SIZE]; char b2[BDEVNAME_SIZE]; char b2[BDEVNAME_SIZE]; struct r0conf *conf = kzalloc(sizeof(*conf), GFP_KERNEL); struct r0conf *conf = kzalloc(sizeof(*conf), GFP_KERNEL); bool discard_supported = false; if (!conf) if (!conf) return -ENOMEM; return -ENOMEM; Loading Loading @@ -195,6 +196,9 @@ static int create_strip_zones(struct mddev *mddev, struct r0conf **private_conf) if (!smallest || (rdev1->sectors < smallest->sectors)) if (!smallest || (rdev1->sectors < smallest->sectors)) smallest = rdev1; smallest = rdev1; cnt++; cnt++; if (blk_queue_discard(bdev_get_queue(rdev1->bdev))) discard_supported = true; } } if (cnt != mddev->raid_disks) { if (cnt != mddev->raid_disks) { printk(KERN_ERR "md/raid0:%s: too few disks (%d of %d) - " printk(KERN_ERR "md/raid0:%s: too few disks (%d of %d) - " Loading Loading @@ -272,6 +276,11 @@ static int create_strip_zones(struct mddev *mddev, struct r0conf **private_conf) blk_queue_io_opt(mddev->queue, blk_queue_io_opt(mddev->queue, (mddev->chunk_sectors << 9) * mddev->raid_disks); (mddev->chunk_sectors << 9) * mddev->raid_disks); if (!discard_supported) queue_flag_clear_unlocked(QUEUE_FLAG_DISCARD, mddev->queue); else queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, mddev->queue); pr_debug("md/raid0:%s: done.\n", mdname(mddev)); pr_debug("md/raid0:%s: done.\n", mdname(mddev)); *private_conf = conf; *private_conf = conf; Loading Loading @@ -423,6 +432,7 @@ static int raid0_run(struct mddev *mddev) return -EINVAL; return -EINVAL; blk_queue_max_hw_sectors(mddev->queue, mddev->chunk_sectors); blk_queue_max_hw_sectors(mddev->queue, mddev->chunk_sectors); blk_queue_max_write_same_sectors(mddev->queue, mddev->chunk_sectors); blk_queue_max_write_same_sectors(mddev->queue, mddev->chunk_sectors); blk_queue_max_discard_sectors(mddev->queue, mddev->chunk_sectors); /* if private is not null, we are here after takeover */ /* if private is not null, we are here after takeover */ if (mddev->private == NULL) { if (mddev->private == NULL) { Loading Loading @@ -510,7 +520,7 @@ static void raid0_make_request(struct mddev *mddev, struct bio *bio) sector_t sector = bio->bi_sector; sector_t sector = bio->bi_sector; struct bio_pair *bp; struct bio_pair *bp; /* Sanity check -- queue functions should prevent this happening */ /* Sanity check -- queue functions should prevent this happening */ if (bio->bi_vcnt != 1 || if ((bio->bi_vcnt != 1 && bio->bi_vcnt != 0) || bio->bi_idx != 0) bio->bi_idx != 0) goto bad_map; goto bad_map; /* This is a one page bio that upper layers /* This is a one page bio that upper layers Loading @@ -536,6 +546,13 @@ static void raid0_make_request(struct mddev *mddev, struct bio *bio) bio->bi_sector = sector_offset + zone->dev_start + bio->bi_sector = sector_offset + zone->dev_start + tmp_dev->data_offset; tmp_dev->data_offset; if (unlikely((bio->bi_rw & REQ_DISCARD) && !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) { /* Just ignore it */ bio_endio(bio, 0); return; } generic_make_request(bio); generic_make_request(bio); return; return; Loading