Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 4d8bb537 authored by Mike Travis's avatar Mike Travis Committed by Ingo Molnar
Browse files

cpumask: use cpumask_var_t in acpi-cpufreq.c



Impact: cleanup, reduce stack usage, use new cpumask API.

Replace the cpumask_t in struct drv_cmd with a cpumask_var_t.  Remove unneeded
online_policy_cpus cpumask_t in acpi_cpufreq_target.  Update refs to use
new cpumask API.

Signed-off-by: default avatarMike Travis <travis@sgi.com>
Acked-by: default avatarRusty Russell <rusty@rustcorp.com.au>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent c74f31c0
Loading
Loading
Loading
Loading
+29 −29
Original line number Diff line number Diff line
@@ -145,7 +145,7 @@ typedef union {

struct drv_cmd {
	unsigned int type;
	cpumask_t mask;
	cpumask_var_t mask;
	drv_addr_union addr;
	u32 val;
};
@@ -193,7 +193,7 @@ static void drv_read(struct drv_cmd *cmd)
	cpumask_t saved_mask = current->cpus_allowed;
	cmd->val = 0;

	set_cpus_allowed_ptr(current, &cmd->mask);
	set_cpus_allowed_ptr(current, cmd->mask);
	do_drv_read(cmd);
	set_cpus_allowed_ptr(current, &saved_mask);
}
@@ -203,8 +203,8 @@ static void drv_write(struct drv_cmd *cmd)
	cpumask_t saved_mask = current->cpus_allowed;
	unsigned int i;

	for_each_cpu_mask_nr(i, cmd->mask) {
		set_cpus_allowed_ptr(current, &cpumask_of_cpu(i));
	for_each_cpu(i, cmd->mask) {
		set_cpus_allowed_ptr(current, cpumask_of(i));
		do_drv_write(cmd);
	}

@@ -212,22 +212,22 @@ static void drv_write(struct drv_cmd *cmd)
	return;
}

static u32 get_cur_val(const cpumask_t *mask)
static u32 get_cur_val(const struct cpumask *mask)
{
	struct acpi_processor_performance *perf;
	struct drv_cmd cmd;

	if (unlikely(cpus_empty(*mask)))
	if (unlikely(cpumask_empty(mask)))
		return 0;

	switch (per_cpu(drv_data, first_cpu(*mask))->cpu_feature) {
	switch (per_cpu(drv_data, cpumask_first(mask))->cpu_feature) {
	case SYSTEM_INTEL_MSR_CAPABLE:
		cmd.type = SYSTEM_INTEL_MSR_CAPABLE;
		cmd.addr.msr.reg = MSR_IA32_PERF_STATUS;
		break;
	case SYSTEM_IO_CAPABLE:
		cmd.type = SYSTEM_IO_CAPABLE;
		perf = per_cpu(drv_data, first_cpu(*mask))->acpi_data;
		perf = per_cpu(drv_data, cpumask_first(mask))->acpi_data;
		cmd.addr.io.port = perf->control_register.address;
		cmd.addr.io.bit_width = perf->control_register.bit_width;
		break;
@@ -235,7 +235,7 @@ static u32 get_cur_val(const cpumask_t *mask)
		return 0;
	}

	cmd.mask = *mask;
	cpumask_copy(cmd.mask, mask);

	drv_read(&cmd);

@@ -386,7 +386,6 @@ static int acpi_cpufreq_target(struct cpufreq_policy *policy,
	struct acpi_cpufreq_data *data = per_cpu(drv_data, policy->cpu);
	struct acpi_processor_performance *perf;
	struct cpufreq_freqs freqs;
	cpumask_t online_policy_cpus;
	struct drv_cmd cmd;
	unsigned int next_state = 0; /* Index into freq_table */
	unsigned int next_perf_state = 0; /* Index into perf table */
@@ -401,20 +400,18 @@ static int acpi_cpufreq_target(struct cpufreq_policy *policy,
		return -ENODEV;
	}

	if (unlikely(!alloc_cpumask_var(&cmd.mask, GFP_KERNEL)))
		return -ENOMEM;

	perf = data->acpi_data;
	result = cpufreq_frequency_table_target(policy,
						data->freq_table,
						target_freq,
						relation, &next_state);
	if (unlikely(result))
		return -ENODEV;

#ifdef CONFIG_HOTPLUG_CPU
	/* cpufreq holds the hotplug lock, so we are safe from here on */
	cpumask_and(&online_policy_cpus, cpu_online_mask, policy->cpus);
#else
	online_policy_cpus = policy->cpus;
#endif
	if (unlikely(result)) {
		result = -ENODEV;
		goto out;
	}

	next_perf_state = data->freq_table[next_state].index;
	if (perf->state == next_perf_state) {
@@ -425,7 +422,7 @@ static int acpi_cpufreq_target(struct cpufreq_policy *policy,
		} else {
			dprintk("Already at target state (P%d)\n",
				next_perf_state);
			return 0;
			goto out;
		}
	}

@@ -444,19 +441,19 @@ static int acpi_cpufreq_target(struct cpufreq_policy *policy,
		cmd.val = (u32) perf->states[next_perf_state].control;
		break;
	default:
		return -ENODEV;
		result = -ENODEV;
		goto out;
	}

	cpus_clear(cmd.mask);

	/* cpufreq holds the hotplug lock, so we are safe from here on */
	if (policy->shared_type != CPUFREQ_SHARED_TYPE_ANY)
		cmd.mask = online_policy_cpus;
		cpumask_and(cmd.mask, cpu_online_mask, policy->cpus);
	else
		cpu_set(policy->cpu, cmd.mask);
		cpumask_copy(cmd.mask, cpumask_of(policy->cpu));

	freqs.old = perf->states[perf->state].core_frequency * 1000;
	freqs.new = data->freq_table[next_state].frequency;
	for_each_cpu_mask_nr(i, cmd.mask) {
	for_each_cpu(i, cmd.mask) {
		freqs.cpu = i;
		cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE);
	}
@@ -464,19 +461,22 @@ static int acpi_cpufreq_target(struct cpufreq_policy *policy,
	drv_write(&cmd);

	if (acpi_pstate_strict) {
		if (!check_freqs(&cmd.mask, freqs.new, data)) {
		if (!check_freqs(cmd.mask, freqs.new, data)) {
			dprintk("acpi_cpufreq_target failed (%d)\n",
				policy->cpu);
			return -EAGAIN;
			result = -EAGAIN;
			goto out;
		}
	}

	for_each_cpu_mask_nr(i, cmd.mask) {
	for_each_cpu(i, cmd.mask) {
		freqs.cpu = i;
		cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE);
	}
	perf->state = next_perf_state;

out:
	free_cpumask_var(cmd.mask);
	return result;
}