Loading drivers/md/raid5.c +86 −121 Original line number Diff line number Diff line Loading @@ -541,7 +541,7 @@ static void init_stripe(struct stripe_head *sh, sector_t sector, int previous) if (dev->toread || dev->read || dev->towrite || dev->written || test_bit(R5_LOCKED, &dev->flags)) { printk(KERN_ERR "sector=%llx i=%d %p %p %p %p %d\n", pr_err("sector=%llx i=%d %p %p %p %p %d\n", (unsigned long long)sh->sector, i, dev->toread, dev->read, dev->towrite, dev->written, test_bit(R5_LOCKED, &dev->flags)); Loading Loading @@ -2347,10 +2347,8 @@ static void raid5_end_read_request(struct bio * bi) * replacement device. We just fail those on * any error */ printk_ratelimited( KERN_INFO "md/raid:%s: read error corrected" " (%lu sectors at %llu on %s)\n", pr_info_ratelimited( "md/raid:%s: read error corrected (%lu sectors at %llu on %s)\n", mdname(conf->mddev), STRIPE_SECTORS, (unsigned long long)s, bdevname(rdev->bdev, b)); Loading @@ -2370,36 +2368,29 @@ static void raid5_end_read_request(struct bio * bi) clear_bit(R5_UPTODATE, &sh->dev[i].flags); atomic_inc(&rdev->read_errors); if (test_bit(R5_ReadRepl, &sh->dev[i].flags)) printk_ratelimited( KERN_WARNING "md/raid:%s: read error on replacement device " "(sector %llu on %s).\n", pr_warn_ratelimited( "md/raid:%s: read error on replacement device (sector %llu on %s).\n", mdname(conf->mddev), (unsigned long long)s, bdn); else if (conf->mddev->degraded >= conf->max_degraded) { set_bad = 1; printk_ratelimited( KERN_WARNING "md/raid:%s: read error not correctable " "(sector %llu on %s).\n", pr_warn_ratelimited( "md/raid:%s: read error not correctable (sector %llu on %s).\n", mdname(conf->mddev), (unsigned long long)s, bdn); } else if (test_bit(R5_ReWrite, &sh->dev[i].flags)) { /* Oh, no!!! */ set_bad = 1; printk_ratelimited( KERN_WARNING "md/raid:%s: read error NOT corrected!! " "(sector %llu on %s).\n", pr_warn_ratelimited( "md/raid:%s: read error NOT corrected!! (sector %llu on %s).\n", mdname(conf->mddev), (unsigned long long)s, bdn); } else if (atomic_read(&rdev->read_errors) > conf->max_nr_stripes) printk(KERN_WARNING "md/raid:%s: Too many read errors, failing device %s.\n", pr_warn("md/raid:%s: Too many read errors, failing device %s.\n", mdname(conf->mddev), bdn); else retry = 1; Loading Loading @@ -2533,8 +2524,7 @@ static void raid5_error(struct mddev *mddev, struct md_rdev *rdev) set_bit(Faulty, &rdev->flags); set_mask_bits(&mddev->flags, 0, BIT(MD_CHANGE_DEVS) | BIT(MD_CHANGE_PENDING)); printk(KERN_ALERT "md/raid:%s: Disk failure on %s, disabling device.\n" pr_crit("md/raid:%s: Disk failure on %s, disabling device.\n" "md/raid:%s: Operation continuing on %d devices.\n", mdname(mddev), bdevname(rdev->bdev, b), Loading Loading @@ -2861,7 +2851,7 @@ sector_t raid5_compute_blocknr(struct stripe_head *sh, int i, int previous) previous, &dummy1, &sh2); if (check != sh->sector || dummy1 != dd_idx || sh2.pd_idx != sh->pd_idx || sh2.qd_idx != sh->qd_idx) { printk(KERN_ERR "md/raid:%s: compute_blocknr: map not correct\n", pr_warn("md/raid:%s: compute_blocknr: map not correct\n", mdname(conf->mddev)); return 0; } Loading Loading @@ -3782,7 +3772,7 @@ static void handle_parity_checks5(struct r5conf *conf, struct stripe_head *sh, case check_state_compute_run: break; default: printk(KERN_ERR "%s: unknown check_state: %d sector: %llu\n", pr_err("%s: unknown check_state: %d sector: %llu\n", __func__, sh->check_state, (unsigned long long) sh->sector); BUG(); Loading Loading @@ -3946,7 +3936,7 @@ static void handle_parity_checks6(struct r5conf *conf, struct stripe_head *sh, case check_state_compute_run: break; default: printk(KERN_ERR "%s: unknown check_state: %d sector: %llu\n", pr_warn("%s: unknown check_state: %d sector: %llu\n", __func__, sh->check_state, (unsigned long long) sh->sector); BUG(); Loading Loading @@ -6385,7 +6375,7 @@ static int raid456_cpu_up_prepare(unsigned int cpu, struct hlist_node *node) struct raid5_percpu *percpu = per_cpu_ptr(conf->percpu, cpu); if (alloc_scratch_buffer(conf, percpu)) { pr_err("%s: failed memory allocation for cpu%u\n", pr_warn("%s: failed memory allocation for cpu%u\n", __func__, cpu); return -ENOMEM; } Loading Loading @@ -6456,7 +6446,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) if (mddev->new_level != 5 && mddev->new_level != 4 && mddev->new_level != 6) { printk(KERN_ERR "md/raid:%s: raid level not set to 4/5/6 (%d)\n", pr_warn("md/raid:%s: raid level not set to 4/5/6 (%d)\n", mdname(mddev), mddev->new_level); return ERR_PTR(-EIO); } Loading @@ -6464,12 +6454,12 @@ static struct r5conf *setup_conf(struct mddev *mddev) && !algorithm_valid_raid5(mddev->new_layout)) || (mddev->new_level == 6 && !algorithm_valid_raid6(mddev->new_layout))) { printk(KERN_ERR "md/raid:%s: layout %d not supported\n", pr_warn("md/raid:%s: layout %d not supported\n", mdname(mddev), mddev->new_layout); return ERR_PTR(-EIO); } if (mddev->new_level == 6 && mddev->raid_disks < 4) { printk(KERN_ERR "md/raid:%s: not enough configured devices (%d, minimum 4)\n", pr_warn("md/raid:%s: not enough configured devices (%d, minimum 4)\n", mdname(mddev), mddev->raid_disks); return ERR_PTR(-EINVAL); } Loading @@ -6477,7 +6467,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) if (!mddev->new_chunk_sectors || (mddev->new_chunk_sectors << 9) % PAGE_SIZE || !is_power_of_2(mddev->new_chunk_sectors)) { printk(KERN_ERR "md/raid:%s: invalid chunk size %d\n", pr_warn("md/raid:%s: invalid chunk size %d\n", mdname(mddev), mddev->new_chunk_sectors << 9); return ERR_PTR(-EINVAL); } Loading Loading @@ -6569,8 +6559,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) if (test_bit(In_sync, &rdev->flags)) { char b[BDEVNAME_SIZE]; printk(KERN_INFO "md/raid:%s: device %s operational as raid" " disk %d\n", pr_info("md/raid:%s: device %s operational as raid disk %d\n", mdname(mddev), bdevname(rdev->bdev, b), raid_disk); } else if (rdev->saved_raid_disk != raid_disk) /* Cannot rely on bitmap to complete recovery */ Loading Loading @@ -6605,21 +6594,18 @@ static struct r5conf *setup_conf(struct mddev *mddev) ((mddev->new_chunk_sectors << 9) / STRIPE_SIZE) * 4); conf->min_nr_stripes = max(NR_STRIPES, stripes); if (conf->min_nr_stripes != NR_STRIPES) printk(KERN_INFO "md/raid:%s: force stripe size %d for reshape\n", pr_info("md/raid:%s: force stripe size %d for reshape\n", mdname(mddev), conf->min_nr_stripes); } memory = conf->min_nr_stripes * (sizeof(struct stripe_head) + max_disks * ((sizeof(struct bio) + PAGE_SIZE))) / 1024; atomic_set(&conf->empty_inactive_list_nr, NR_STRIPE_HASH_LOCKS); if (grow_stripes(conf, conf->min_nr_stripes)) { printk(KERN_ERR "md/raid:%s: couldn't allocate %dkB for buffers\n", pr_warn("md/raid:%s: couldn't allocate %dkB for buffers\n", mdname(mddev), memory); goto abort; } else printk(KERN_INFO "md/raid:%s: allocated %dkB\n", mdname(mddev), memory); pr_debug("md/raid:%s: allocated %dkB\n", mdname(mddev), memory); /* * Losing a stripe head costs more than the time to refill it, * it reduces the queue depth and so can hurt throughput. Loading @@ -6631,8 +6617,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) conf->shrinker.batch = 128; conf->shrinker.flags = 0; if (register_shrinker(&conf->shrinker)) { printk(KERN_ERR "md/raid:%s: couldn't register shrinker.\n", pr_warn("md/raid:%s: couldn't register shrinker.\n", mdname(mddev)); goto abort; } Loading @@ -6640,8 +6625,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) sprintf(pers_name, "raid%d", mddev->new_level); conf->thread = md_register_thread(raid5d, mddev, pers_name); if (!conf->thread) { printk(KERN_ERR "md/raid:%s: couldn't allocate thread.\n", pr_warn("md/raid:%s: couldn't allocate thread.\n", mdname(mddev)); goto abort; } Loading Loading @@ -6695,8 +6679,7 @@ static int raid5_run(struct mddev *mddev) int first = 1; if (mddev->recovery_cp != MaxSector) printk(KERN_NOTICE "md/raid:%s: not clean" " -- starting background reconstruction\n", pr_notice("md/raid:%s: not clean -- starting background reconstruction\n", mdname(mddev)); rdev_for_each(rdev, mddev) { Loading Loading @@ -6740,14 +6723,13 @@ static int raid5_run(struct mddev *mddev) int new_data_disks; if (journal_dev) { printk(KERN_ERR "md/raid:%s: don't support reshape with journal - aborting.\n", pr_warn("md/raid:%s: don't support reshape with journal - aborting.\n", mdname(mddev)); return -EINVAL; } if (mddev->new_level != mddev->level) { printk(KERN_ERR "md/raid:%s: unsupported reshape " "required - aborting.\n", pr_warn("md/raid:%s: unsupported reshape required - aborting.\n", mdname(mddev)); return -EINVAL; } Loading @@ -6763,8 +6745,8 @@ static int raid5_run(struct mddev *mddev) chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); new_data_disks = mddev->raid_disks - max_degraded; if (sector_div(here_new, chunk_sectors * new_data_disks)) { printk(KERN_ERR "md/raid:%s: reshape_position not " "on a stripe boundary\n", mdname(mddev)); pr_warn("md/raid:%s: reshape_position not on a stripe boundary\n", mdname(mddev)); return -EINVAL; } reshape_offset = here_new * chunk_sectors; Loading @@ -6785,9 +6767,7 @@ static int raid5_run(struct mddev *mddev) abs(min_offset_diff) >= mddev->new_chunk_sectors) /* not really in-place - so OK */; else if (mddev->ro == 0) { printk(KERN_ERR "md/raid:%s: in-place reshape " "must be started in read-only mode " "- aborting\n", pr_warn("md/raid:%s: in-place reshape must be started in read-only mode - aborting\n", mdname(mddev)); return -EINVAL; } Loading @@ -6797,13 +6777,11 @@ static int raid5_run(struct mddev *mddev) : (here_new * chunk_sectors >= here_old * chunk_sectors + (-min_offset_diff))) { /* Reading from the same stripe as writing to - bad */ printk(KERN_ERR "md/raid:%s: reshape_position too early for " "auto-recovery - aborting.\n", pr_warn("md/raid:%s: reshape_position too early for auto-recovery - aborting.\n", mdname(mddev)); return -EINVAL; } printk(KERN_INFO "md/raid:%s: reshape will continue\n", mdname(mddev)); pr_debug("md/raid:%s: reshape will continue\n", mdname(mddev)); /* OK, we should be able to continue; */ } else { BUG_ON(mddev->level != mddev->new_level); Loading @@ -6822,7 +6800,7 @@ static int raid5_run(struct mddev *mddev) if (test_bit(MD_HAS_JOURNAL, &mddev->flags)) { if (!journal_dev) { pr_err("md/raid:%s: journal disk is missing, force array readonly\n", pr_warn("md/raid:%s: journal disk is missing, force array readonly\n", mdname(mddev)); mddev->ro = 1; set_disk_ro(mddev->gendisk, 1); Loading Loading @@ -6850,8 +6828,7 @@ static int raid5_run(struct mddev *mddev) if (conf->disks[i].replacement && conf->reshape_progress != MaxSector) { /* replacements and reshape simply do not mix. */ printk(KERN_ERR "md: cannot handle concurrent " "replacement and reshape.\n"); pr_warn("md: cannot handle concurrent replacement and reshape.\n"); goto abort; } if (test_bit(In_sync, &rdev->flags)) { Loading Loading @@ -6893,8 +6870,7 @@ static int raid5_run(struct mddev *mddev) mddev->degraded = calc_degraded(conf); if (has_failed(conf)) { printk(KERN_ERR "md/raid:%s: not enough operational devices" " (%d/%d failed)\n", pr_crit("md/raid:%s: not enough operational devices (%d/%d failed)\n", mdname(mddev), mddev->degraded, conf->raid_disks); goto abort; } Loading @@ -6906,29 +6882,19 @@ static int raid5_run(struct mddev *mddev) if (mddev->degraded > dirty_parity_disks && mddev->recovery_cp != MaxSector) { if (mddev->ok_start_degraded) printk(KERN_WARNING "md/raid:%s: starting dirty degraded array" " - data corruption possible.\n", pr_crit("md/raid:%s: starting dirty degraded array - data corruption possible.\n", mdname(mddev)); else { printk(KERN_ERR "md/raid:%s: cannot start dirty degraded array.\n", pr_crit("md/raid:%s: cannot start dirty degraded array.\n", mdname(mddev)); goto abort; } } if (mddev->degraded == 0) printk(KERN_INFO "md/raid:%s: raid level %d active with %d out of %d" " devices, algorithm %d\n", mdname(mddev), conf->level, pr_info("md/raid:%s: raid level %d active with %d out of %d devices, algorithm %d\n", mdname(mddev), conf->level, mddev->raid_disks-mddev->degraded, mddev->raid_disks, mddev->new_layout); else printk(KERN_ALERT "md/raid:%s: raid level %d active with %d" " out of %d devices, algorithm %d\n", mdname(mddev), conf->level, mddev->raid_disks - mddev->degraded, mddev->raid_disks, mddev->new_layout); print_raid5_conf(conf); Loading @@ -6948,8 +6914,7 @@ static int raid5_run(struct mddev *mddev) mddev->to_remove = NULL; else if (mddev->kobj.sd && sysfs_create_group(&mddev->kobj, &raid5_attrs_group)) printk(KERN_WARNING "raid5: failed to create sysfs attributes for %s\n", pr_warn("raid5: failed to create sysfs attributes for %s\n", mdname(mddev)); md_set_array_sectors(mddev, raid5_size(mddev, 0, 0)); Loading Loading @@ -7038,7 +7003,7 @@ static int raid5_run(struct mddev *mddev) if (journal_dev) { char b[BDEVNAME_SIZE]; printk(KERN_INFO"md/raid:%s: using device %s as journal\n", pr_debug("md/raid:%s: using device %s as journal\n", mdname(mddev), bdevname(journal_dev->bdev, b)); r5l_init_log(conf, journal_dev); } Loading @@ -7049,7 +7014,7 @@ static int raid5_run(struct mddev *mddev) print_raid5_conf(conf); free_conf(conf); mddev->private = NULL; printk(KERN_ALERT "md/raid:%s: failed to run raid set.\n", mdname(mddev)); pr_warn("md/raid:%s: failed to run raid set.\n", mdname(mddev)); return -EIO; } Loading Loading @@ -7083,12 +7048,12 @@ static void print_raid5_conf (struct r5conf *conf) int i; struct disk_info *tmp; printk(KERN_DEBUG "RAID conf printout:\n"); pr_debug("RAID conf printout:\n"); if (!conf) { printk("(conf==NULL)\n"); pr_debug("(conf==NULL)\n"); return; } printk(KERN_DEBUG " --- level:%d rd:%d wd:%d\n", conf->level, pr_debug(" --- level:%d rd:%d wd:%d\n", conf->level, conf->raid_disks, conf->raid_disks - conf->mddev->degraded); Loading @@ -7096,7 +7061,7 @@ static void print_raid5_conf (struct r5conf *conf) char b[BDEVNAME_SIZE]; tmp = conf->disks + i; if (tmp->rdev) printk(KERN_DEBUG " disk %d, o:%d, dev:%s\n", pr_debug(" disk %d, o:%d, dev:%s\n", i, !test_bit(Faulty, &tmp->rdev->flags), bdevname(tmp->rdev->bdev, b)); } Loading Loading @@ -7244,7 +7209,7 @@ static int raid5_add_disk(struct mddev *mddev, struct md_rdev *rdev) * write requests running. We should be safe */ r5l_init_log(conf, rdev); printk(KERN_INFO"md/raid:%s: using device %s as journal\n", pr_debug("md/raid:%s: using device %s as journal\n", mdname(mddev), bdevname(rdev->bdev, b)); return 0; } Loading Loading @@ -7349,7 +7314,7 @@ static int check_stripe_cache(struct mddev *mddev) > conf->min_nr_stripes || ((mddev->new_chunk_sectors << 9) / STRIPE_SIZE) * 4 > conf->min_nr_stripes) { printk(KERN_WARNING "md/raid:%s: reshape: not enough stripes. Needed %lu\n", pr_warn("md/raid:%s: reshape: not enough stripes. Needed %lu\n", mdname(mddev), ((max(mddev->chunk_sectors, mddev->new_chunk_sectors) << 9) / STRIPE_SIZE)*4); Loading Loading @@ -7433,8 +7398,8 @@ static int raid5_start_reshape(struct mddev *mddev) */ if (raid5_size(mddev, 0, conf->raid_disks + mddev->delta_disks) < mddev->array_sectors) { printk(KERN_ERR "md/raid:%s: array size must be reduced " "before number of disks\n", mdname(mddev)); pr_warn("md/raid:%s: array size must be reduced before number of disks\n", mdname(mddev)); return -EINVAL; } Loading Loading @@ -7652,7 +7617,7 @@ static void *raid45_takeover_raid0(struct mddev *mddev, int level) /* for raid0 takeover only one zone is supported */ if (raid0_conf->nr_strip_zones > 1) { printk(KERN_ERR "md/raid:%s: cannot takeover raid0 with more than one zone.\n", pr_warn("md/raid:%s: cannot takeover raid0 with more than one zone.\n", mdname(mddev)); return ERR_PTR(-EINVAL); } Loading Loading
drivers/md/raid5.c +86 −121 Original line number Diff line number Diff line Loading @@ -541,7 +541,7 @@ static void init_stripe(struct stripe_head *sh, sector_t sector, int previous) if (dev->toread || dev->read || dev->towrite || dev->written || test_bit(R5_LOCKED, &dev->flags)) { printk(KERN_ERR "sector=%llx i=%d %p %p %p %p %d\n", pr_err("sector=%llx i=%d %p %p %p %p %d\n", (unsigned long long)sh->sector, i, dev->toread, dev->read, dev->towrite, dev->written, test_bit(R5_LOCKED, &dev->flags)); Loading Loading @@ -2347,10 +2347,8 @@ static void raid5_end_read_request(struct bio * bi) * replacement device. We just fail those on * any error */ printk_ratelimited( KERN_INFO "md/raid:%s: read error corrected" " (%lu sectors at %llu on %s)\n", pr_info_ratelimited( "md/raid:%s: read error corrected (%lu sectors at %llu on %s)\n", mdname(conf->mddev), STRIPE_SECTORS, (unsigned long long)s, bdevname(rdev->bdev, b)); Loading @@ -2370,36 +2368,29 @@ static void raid5_end_read_request(struct bio * bi) clear_bit(R5_UPTODATE, &sh->dev[i].flags); atomic_inc(&rdev->read_errors); if (test_bit(R5_ReadRepl, &sh->dev[i].flags)) printk_ratelimited( KERN_WARNING "md/raid:%s: read error on replacement device " "(sector %llu on %s).\n", pr_warn_ratelimited( "md/raid:%s: read error on replacement device (sector %llu on %s).\n", mdname(conf->mddev), (unsigned long long)s, bdn); else if (conf->mddev->degraded >= conf->max_degraded) { set_bad = 1; printk_ratelimited( KERN_WARNING "md/raid:%s: read error not correctable " "(sector %llu on %s).\n", pr_warn_ratelimited( "md/raid:%s: read error not correctable (sector %llu on %s).\n", mdname(conf->mddev), (unsigned long long)s, bdn); } else if (test_bit(R5_ReWrite, &sh->dev[i].flags)) { /* Oh, no!!! */ set_bad = 1; printk_ratelimited( KERN_WARNING "md/raid:%s: read error NOT corrected!! " "(sector %llu on %s).\n", pr_warn_ratelimited( "md/raid:%s: read error NOT corrected!! (sector %llu on %s).\n", mdname(conf->mddev), (unsigned long long)s, bdn); } else if (atomic_read(&rdev->read_errors) > conf->max_nr_stripes) printk(KERN_WARNING "md/raid:%s: Too many read errors, failing device %s.\n", pr_warn("md/raid:%s: Too many read errors, failing device %s.\n", mdname(conf->mddev), bdn); else retry = 1; Loading Loading @@ -2533,8 +2524,7 @@ static void raid5_error(struct mddev *mddev, struct md_rdev *rdev) set_bit(Faulty, &rdev->flags); set_mask_bits(&mddev->flags, 0, BIT(MD_CHANGE_DEVS) | BIT(MD_CHANGE_PENDING)); printk(KERN_ALERT "md/raid:%s: Disk failure on %s, disabling device.\n" pr_crit("md/raid:%s: Disk failure on %s, disabling device.\n" "md/raid:%s: Operation continuing on %d devices.\n", mdname(mddev), bdevname(rdev->bdev, b), Loading Loading @@ -2861,7 +2851,7 @@ sector_t raid5_compute_blocknr(struct stripe_head *sh, int i, int previous) previous, &dummy1, &sh2); if (check != sh->sector || dummy1 != dd_idx || sh2.pd_idx != sh->pd_idx || sh2.qd_idx != sh->qd_idx) { printk(KERN_ERR "md/raid:%s: compute_blocknr: map not correct\n", pr_warn("md/raid:%s: compute_blocknr: map not correct\n", mdname(conf->mddev)); return 0; } Loading Loading @@ -3782,7 +3772,7 @@ static void handle_parity_checks5(struct r5conf *conf, struct stripe_head *sh, case check_state_compute_run: break; default: printk(KERN_ERR "%s: unknown check_state: %d sector: %llu\n", pr_err("%s: unknown check_state: %d sector: %llu\n", __func__, sh->check_state, (unsigned long long) sh->sector); BUG(); Loading Loading @@ -3946,7 +3936,7 @@ static void handle_parity_checks6(struct r5conf *conf, struct stripe_head *sh, case check_state_compute_run: break; default: printk(KERN_ERR "%s: unknown check_state: %d sector: %llu\n", pr_warn("%s: unknown check_state: %d sector: %llu\n", __func__, sh->check_state, (unsigned long long) sh->sector); BUG(); Loading Loading @@ -6385,7 +6375,7 @@ static int raid456_cpu_up_prepare(unsigned int cpu, struct hlist_node *node) struct raid5_percpu *percpu = per_cpu_ptr(conf->percpu, cpu); if (alloc_scratch_buffer(conf, percpu)) { pr_err("%s: failed memory allocation for cpu%u\n", pr_warn("%s: failed memory allocation for cpu%u\n", __func__, cpu); return -ENOMEM; } Loading Loading @@ -6456,7 +6446,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) if (mddev->new_level != 5 && mddev->new_level != 4 && mddev->new_level != 6) { printk(KERN_ERR "md/raid:%s: raid level not set to 4/5/6 (%d)\n", pr_warn("md/raid:%s: raid level not set to 4/5/6 (%d)\n", mdname(mddev), mddev->new_level); return ERR_PTR(-EIO); } Loading @@ -6464,12 +6454,12 @@ static struct r5conf *setup_conf(struct mddev *mddev) && !algorithm_valid_raid5(mddev->new_layout)) || (mddev->new_level == 6 && !algorithm_valid_raid6(mddev->new_layout))) { printk(KERN_ERR "md/raid:%s: layout %d not supported\n", pr_warn("md/raid:%s: layout %d not supported\n", mdname(mddev), mddev->new_layout); return ERR_PTR(-EIO); } if (mddev->new_level == 6 && mddev->raid_disks < 4) { printk(KERN_ERR "md/raid:%s: not enough configured devices (%d, minimum 4)\n", pr_warn("md/raid:%s: not enough configured devices (%d, minimum 4)\n", mdname(mddev), mddev->raid_disks); return ERR_PTR(-EINVAL); } Loading @@ -6477,7 +6467,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) if (!mddev->new_chunk_sectors || (mddev->new_chunk_sectors << 9) % PAGE_SIZE || !is_power_of_2(mddev->new_chunk_sectors)) { printk(KERN_ERR "md/raid:%s: invalid chunk size %d\n", pr_warn("md/raid:%s: invalid chunk size %d\n", mdname(mddev), mddev->new_chunk_sectors << 9); return ERR_PTR(-EINVAL); } Loading Loading @@ -6569,8 +6559,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) if (test_bit(In_sync, &rdev->flags)) { char b[BDEVNAME_SIZE]; printk(KERN_INFO "md/raid:%s: device %s operational as raid" " disk %d\n", pr_info("md/raid:%s: device %s operational as raid disk %d\n", mdname(mddev), bdevname(rdev->bdev, b), raid_disk); } else if (rdev->saved_raid_disk != raid_disk) /* Cannot rely on bitmap to complete recovery */ Loading Loading @@ -6605,21 +6594,18 @@ static struct r5conf *setup_conf(struct mddev *mddev) ((mddev->new_chunk_sectors << 9) / STRIPE_SIZE) * 4); conf->min_nr_stripes = max(NR_STRIPES, stripes); if (conf->min_nr_stripes != NR_STRIPES) printk(KERN_INFO "md/raid:%s: force stripe size %d for reshape\n", pr_info("md/raid:%s: force stripe size %d for reshape\n", mdname(mddev), conf->min_nr_stripes); } memory = conf->min_nr_stripes * (sizeof(struct stripe_head) + max_disks * ((sizeof(struct bio) + PAGE_SIZE))) / 1024; atomic_set(&conf->empty_inactive_list_nr, NR_STRIPE_HASH_LOCKS); if (grow_stripes(conf, conf->min_nr_stripes)) { printk(KERN_ERR "md/raid:%s: couldn't allocate %dkB for buffers\n", pr_warn("md/raid:%s: couldn't allocate %dkB for buffers\n", mdname(mddev), memory); goto abort; } else printk(KERN_INFO "md/raid:%s: allocated %dkB\n", mdname(mddev), memory); pr_debug("md/raid:%s: allocated %dkB\n", mdname(mddev), memory); /* * Losing a stripe head costs more than the time to refill it, * it reduces the queue depth and so can hurt throughput. Loading @@ -6631,8 +6617,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) conf->shrinker.batch = 128; conf->shrinker.flags = 0; if (register_shrinker(&conf->shrinker)) { printk(KERN_ERR "md/raid:%s: couldn't register shrinker.\n", pr_warn("md/raid:%s: couldn't register shrinker.\n", mdname(mddev)); goto abort; } Loading @@ -6640,8 +6625,7 @@ static struct r5conf *setup_conf(struct mddev *mddev) sprintf(pers_name, "raid%d", mddev->new_level); conf->thread = md_register_thread(raid5d, mddev, pers_name); if (!conf->thread) { printk(KERN_ERR "md/raid:%s: couldn't allocate thread.\n", pr_warn("md/raid:%s: couldn't allocate thread.\n", mdname(mddev)); goto abort; } Loading Loading @@ -6695,8 +6679,7 @@ static int raid5_run(struct mddev *mddev) int first = 1; if (mddev->recovery_cp != MaxSector) printk(KERN_NOTICE "md/raid:%s: not clean" " -- starting background reconstruction\n", pr_notice("md/raid:%s: not clean -- starting background reconstruction\n", mdname(mddev)); rdev_for_each(rdev, mddev) { Loading Loading @@ -6740,14 +6723,13 @@ static int raid5_run(struct mddev *mddev) int new_data_disks; if (journal_dev) { printk(KERN_ERR "md/raid:%s: don't support reshape with journal - aborting.\n", pr_warn("md/raid:%s: don't support reshape with journal - aborting.\n", mdname(mddev)); return -EINVAL; } if (mddev->new_level != mddev->level) { printk(KERN_ERR "md/raid:%s: unsupported reshape " "required - aborting.\n", pr_warn("md/raid:%s: unsupported reshape required - aborting.\n", mdname(mddev)); return -EINVAL; } Loading @@ -6763,8 +6745,8 @@ static int raid5_run(struct mddev *mddev) chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); new_data_disks = mddev->raid_disks - max_degraded; if (sector_div(here_new, chunk_sectors * new_data_disks)) { printk(KERN_ERR "md/raid:%s: reshape_position not " "on a stripe boundary\n", mdname(mddev)); pr_warn("md/raid:%s: reshape_position not on a stripe boundary\n", mdname(mddev)); return -EINVAL; } reshape_offset = here_new * chunk_sectors; Loading @@ -6785,9 +6767,7 @@ static int raid5_run(struct mddev *mddev) abs(min_offset_diff) >= mddev->new_chunk_sectors) /* not really in-place - so OK */; else if (mddev->ro == 0) { printk(KERN_ERR "md/raid:%s: in-place reshape " "must be started in read-only mode " "- aborting\n", pr_warn("md/raid:%s: in-place reshape must be started in read-only mode - aborting\n", mdname(mddev)); return -EINVAL; } Loading @@ -6797,13 +6777,11 @@ static int raid5_run(struct mddev *mddev) : (here_new * chunk_sectors >= here_old * chunk_sectors + (-min_offset_diff))) { /* Reading from the same stripe as writing to - bad */ printk(KERN_ERR "md/raid:%s: reshape_position too early for " "auto-recovery - aborting.\n", pr_warn("md/raid:%s: reshape_position too early for auto-recovery - aborting.\n", mdname(mddev)); return -EINVAL; } printk(KERN_INFO "md/raid:%s: reshape will continue\n", mdname(mddev)); pr_debug("md/raid:%s: reshape will continue\n", mdname(mddev)); /* OK, we should be able to continue; */ } else { BUG_ON(mddev->level != mddev->new_level); Loading @@ -6822,7 +6800,7 @@ static int raid5_run(struct mddev *mddev) if (test_bit(MD_HAS_JOURNAL, &mddev->flags)) { if (!journal_dev) { pr_err("md/raid:%s: journal disk is missing, force array readonly\n", pr_warn("md/raid:%s: journal disk is missing, force array readonly\n", mdname(mddev)); mddev->ro = 1; set_disk_ro(mddev->gendisk, 1); Loading Loading @@ -6850,8 +6828,7 @@ static int raid5_run(struct mddev *mddev) if (conf->disks[i].replacement && conf->reshape_progress != MaxSector) { /* replacements and reshape simply do not mix. */ printk(KERN_ERR "md: cannot handle concurrent " "replacement and reshape.\n"); pr_warn("md: cannot handle concurrent replacement and reshape.\n"); goto abort; } if (test_bit(In_sync, &rdev->flags)) { Loading Loading @@ -6893,8 +6870,7 @@ static int raid5_run(struct mddev *mddev) mddev->degraded = calc_degraded(conf); if (has_failed(conf)) { printk(KERN_ERR "md/raid:%s: not enough operational devices" " (%d/%d failed)\n", pr_crit("md/raid:%s: not enough operational devices (%d/%d failed)\n", mdname(mddev), mddev->degraded, conf->raid_disks); goto abort; } Loading @@ -6906,29 +6882,19 @@ static int raid5_run(struct mddev *mddev) if (mddev->degraded > dirty_parity_disks && mddev->recovery_cp != MaxSector) { if (mddev->ok_start_degraded) printk(KERN_WARNING "md/raid:%s: starting dirty degraded array" " - data corruption possible.\n", pr_crit("md/raid:%s: starting dirty degraded array - data corruption possible.\n", mdname(mddev)); else { printk(KERN_ERR "md/raid:%s: cannot start dirty degraded array.\n", pr_crit("md/raid:%s: cannot start dirty degraded array.\n", mdname(mddev)); goto abort; } } if (mddev->degraded == 0) printk(KERN_INFO "md/raid:%s: raid level %d active with %d out of %d" " devices, algorithm %d\n", mdname(mddev), conf->level, pr_info("md/raid:%s: raid level %d active with %d out of %d devices, algorithm %d\n", mdname(mddev), conf->level, mddev->raid_disks-mddev->degraded, mddev->raid_disks, mddev->new_layout); else printk(KERN_ALERT "md/raid:%s: raid level %d active with %d" " out of %d devices, algorithm %d\n", mdname(mddev), conf->level, mddev->raid_disks - mddev->degraded, mddev->raid_disks, mddev->new_layout); print_raid5_conf(conf); Loading @@ -6948,8 +6914,7 @@ static int raid5_run(struct mddev *mddev) mddev->to_remove = NULL; else if (mddev->kobj.sd && sysfs_create_group(&mddev->kobj, &raid5_attrs_group)) printk(KERN_WARNING "raid5: failed to create sysfs attributes for %s\n", pr_warn("raid5: failed to create sysfs attributes for %s\n", mdname(mddev)); md_set_array_sectors(mddev, raid5_size(mddev, 0, 0)); Loading Loading @@ -7038,7 +7003,7 @@ static int raid5_run(struct mddev *mddev) if (journal_dev) { char b[BDEVNAME_SIZE]; printk(KERN_INFO"md/raid:%s: using device %s as journal\n", pr_debug("md/raid:%s: using device %s as journal\n", mdname(mddev), bdevname(journal_dev->bdev, b)); r5l_init_log(conf, journal_dev); } Loading @@ -7049,7 +7014,7 @@ static int raid5_run(struct mddev *mddev) print_raid5_conf(conf); free_conf(conf); mddev->private = NULL; printk(KERN_ALERT "md/raid:%s: failed to run raid set.\n", mdname(mddev)); pr_warn("md/raid:%s: failed to run raid set.\n", mdname(mddev)); return -EIO; } Loading Loading @@ -7083,12 +7048,12 @@ static void print_raid5_conf (struct r5conf *conf) int i; struct disk_info *tmp; printk(KERN_DEBUG "RAID conf printout:\n"); pr_debug("RAID conf printout:\n"); if (!conf) { printk("(conf==NULL)\n"); pr_debug("(conf==NULL)\n"); return; } printk(KERN_DEBUG " --- level:%d rd:%d wd:%d\n", conf->level, pr_debug(" --- level:%d rd:%d wd:%d\n", conf->level, conf->raid_disks, conf->raid_disks - conf->mddev->degraded); Loading @@ -7096,7 +7061,7 @@ static void print_raid5_conf (struct r5conf *conf) char b[BDEVNAME_SIZE]; tmp = conf->disks + i; if (tmp->rdev) printk(KERN_DEBUG " disk %d, o:%d, dev:%s\n", pr_debug(" disk %d, o:%d, dev:%s\n", i, !test_bit(Faulty, &tmp->rdev->flags), bdevname(tmp->rdev->bdev, b)); } Loading Loading @@ -7244,7 +7209,7 @@ static int raid5_add_disk(struct mddev *mddev, struct md_rdev *rdev) * write requests running. We should be safe */ r5l_init_log(conf, rdev); printk(KERN_INFO"md/raid:%s: using device %s as journal\n", pr_debug("md/raid:%s: using device %s as journal\n", mdname(mddev), bdevname(rdev->bdev, b)); return 0; } Loading Loading @@ -7349,7 +7314,7 @@ static int check_stripe_cache(struct mddev *mddev) > conf->min_nr_stripes || ((mddev->new_chunk_sectors << 9) / STRIPE_SIZE) * 4 > conf->min_nr_stripes) { printk(KERN_WARNING "md/raid:%s: reshape: not enough stripes. Needed %lu\n", pr_warn("md/raid:%s: reshape: not enough stripes. Needed %lu\n", mdname(mddev), ((max(mddev->chunk_sectors, mddev->new_chunk_sectors) << 9) / STRIPE_SIZE)*4); Loading Loading @@ -7433,8 +7398,8 @@ static int raid5_start_reshape(struct mddev *mddev) */ if (raid5_size(mddev, 0, conf->raid_disks + mddev->delta_disks) < mddev->array_sectors) { printk(KERN_ERR "md/raid:%s: array size must be reduced " "before number of disks\n", mdname(mddev)); pr_warn("md/raid:%s: array size must be reduced before number of disks\n", mdname(mddev)); return -EINVAL; } Loading Loading @@ -7652,7 +7617,7 @@ static void *raid45_takeover_raid0(struct mddev *mddev, int level) /* for raid0 takeover only one zone is supported */ if (raid0_conf->nr_strip_zones > 1) { printk(KERN_ERR "md/raid:%s: cannot takeover raid0 with more than one zone.\n", pr_warn("md/raid:%s: cannot takeover raid0 with more than one zone.\n", mdname(mddev)); return ERR_PTR(-EINVAL); } Loading