Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 73c6f239 authored by Heinz Mauelshagen's avatar Heinz Mauelshagen Committed by Mike Snitzer
Browse files

dm raid: rename variable 'ret' to 'r' to conform to other dm code

parent e83068a5
Loading
Loading
Loading
Loading
+36 −32
Original line number Diff line number Diff line
@@ -225,7 +225,7 @@ static int dev_parms(struct raid_set *rs, char **argv)
	int i;
	int rebuild = 0;
	int metadata_available = 0;
	int ret = 0;
	int r = 0;

	for (i = 0; i < rs->md.raid_disks; i++, argv += 2) {
		rs->dev[i].rdev.raid_disk = i;
@@ -241,12 +241,12 @@ static int dev_parms(struct raid_set *rs, char **argv)
		rs->dev[i].rdev.mddev = &rs->md;

		if (strcmp(argv[0], "-")) {
			ret = dm_get_device(rs->ti, argv[0],
			r = dm_get_device(rs->ti, argv[0],
					    dm_table_get_mode(rs->ti->table),
					    &rs->dev[i].meta_dev);
			rs->ti->error = "RAID metadata device lookup failure";
			if (ret)
				return ret;
			if (r)
				return r;

			rs->dev[i].rdev.sb_page = alloc_page(GFP_KERNEL);
			if (!rs->dev[i].rdev.sb_page)
@@ -267,12 +267,12 @@ static int dev_parms(struct raid_set *rs, char **argv)
			continue;
		}

		ret = dm_get_device(rs->ti, argv[1],
		r = dm_get_device(rs->ti, argv[1],
				    dm_table_get_mode(rs->ti->table),
				    &rs->dev[i].data_dev);
		if (ret) {
		if (r) {
			rs->ti->error = "RAID device lookup failure";
			return ret;
			return r;
		}

		if (rs->dev[i].meta_dev) {
@@ -848,7 +848,7 @@ static void super_sync(struct mddev *mddev, struct md_rdev *rdev)
 */
static int super_load(struct md_rdev *rdev, struct md_rdev *refdev)
{
	int ret;
	int r;
	struct dm_raid_superblock *sb;
	struct dm_raid_superblock *refsb;
	uint64_t events_sb, events_refsb;
@@ -860,9 +860,9 @@ static int super_load(struct md_rdev *rdev, struct md_rdev *refdev)
		return -EINVAL;
	}

	ret = read_disk_sb(rdev, rdev->sb_size);
	if (ret)
		return ret;
	r = read_disk_sb(rdev, rdev->sb_size);
	if (r)
		return r;

	sb = page_address(rdev->sb_page);

@@ -1072,7 +1072,7 @@ static int super_validate(struct raid_set *rs, struct md_rdev *rdev)
 */
static int analyse_superblocks(struct dm_target *ti, struct raid_set *rs)
{
	int ret;
	int r;
	struct raid_dev *dev;
	struct md_rdev *rdev, *tmp, *freshest;
	struct mddev *mddev = &rs->md;
@@ -1097,9 +1097,9 @@ static int analyse_superblocks(struct dm_target *ti, struct raid_set *rs)
		if (!rdev->meta_bdev)
			continue;

		ret = super_load(rdev, freshest);
		r = super_load(rdev, freshest);

		switch (ret) {
		switch (r) {
		case 1:
			freshest = rdev;
			break;
@@ -1207,17 +1207,21 @@ static void configure_discard_support(struct dm_target *ti, struct raid_set *rs)
}

/*
 * Construct a RAID4/5/6 mapping:
 * Construct a RAID0/1/10/4/5/6 mapping:
 * Args:
 *	<raid_type> <#raid_params> <raid_params>		\
 *	<#raid_devs> { <meta_dev1> <dev1> .. <meta_devN> <devN> }
 *      <raid_type> <#raid_params> <raid_params>{0,}    \
 *      <#raid_devs> [<meta_dev1> <dev1>]{1,}
 *
 * <raid_params> varies by <raid_type>.  See 'parse_raid_params' for
 * details on possible <raid_params>.
 *
 * Userspace is free to initialize the metadata devices, hence the superblocks to
 * enforce recreation based on the passed in table parameters.
 *
 */
static int raid_ctr(struct dm_target *ti, unsigned argc, char **argv)
{
	int ret;
	int r;
	struct raid_type *rt;
	unsigned long num_raid_params, num_raid_devs;
	struct raid_set *rs = NULL;
@@ -1267,19 +1271,19 @@ static int raid_ctr(struct dm_target *ti, unsigned argc, char **argv)
	if (IS_ERR(rs))
		return PTR_ERR(rs);

	ret = parse_raid_params(rs, argv, (unsigned)num_raid_params);
	if (ret)
	r = parse_raid_params(rs, argv, (unsigned)num_raid_params);
	if (r)
		goto bad;

	argv += num_raid_params + 1;

	ret = dev_parms(rs, argv);
	if (ret)
	r = dev_parms(rs, argv);
	if (r)
		goto bad;

	rs->md.sync_super = super_sync;
	ret = analyse_superblocks(ti, rs);
	if (ret)
	r = analyse_superblocks(ti, rs);
	if (r)
		goto bad;

	INIT_WORK(&rs->md.event_work, do_table_event);
@@ -1293,18 +1297,18 @@ static int raid_ctr(struct dm_target *ti, unsigned argc, char **argv)

	/* Has to be held on running the array */
	mddev_lock_nointr(&rs->md);
	ret = md_run(&rs->md);
	r = md_run(&rs->md);
	rs->md.in_sync = 0; /* Assume already marked dirty */
	mddev_unlock(&rs->md);

	if (ret) {
	if (r) {
		ti->error = "Fail to run raid array";
		goto bad;
	}

	if (ti->len != rs->md.array_sectors) {
		ti->error = "Array size does not match requested target length";
		ret = -EINVAL;
		r = -EINVAL;
		goto size_mismatch;
	}
	rs->callbacks.congested_fn = raid_is_congested;
@@ -1318,7 +1322,7 @@ static int raid_ctr(struct dm_target *ti, unsigned argc, char **argv)
bad:
	context_free(rs);

	return ret;
	return r;
}

static void raid_dtr(struct dm_target *ti)
@@ -1603,17 +1607,17 @@ static int raid_iterate_devices(struct dm_target *ti,
{
	struct raid_set *rs = ti->private;
	unsigned i;
	int ret = 0;
	int r = 0;

	for (i = 0; !ret && i < rs->md.raid_disks; i++)
	for (i = 0; !r && i < rs->md.raid_disks; i++)
		if (rs->dev[i].data_dev)
			ret = fn(ti,
			r = fn(ti,
				 rs->dev[i].data_dev,
				 0, /* No offset on data devs */
				 rs->md.dev_sectors,
				 data);

	return ret;
	return r;
}

static void raid_io_hints(struct dm_target *ti, struct queue_limits *limits)