Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 1f47b0dd authored by Rafael J. Wysocki's avatar Rafael J. Wysocki
Browse files

Merge branch 'pm-cpufreq'

* pm-cpufreq:
  cpufreq: s5pv210-cpufreq: fix wrong do_div() usage
  MAINTAINERS: update for intel P-state driver
  cpufreq: governor: Quit work-handlers early if governor is stopped
  intel_pstate: decrease number of "HWP enabled" messages
  cpufreq: arm_big_little: fix frequency check when bL switcher is active
parents 1ede53f7 d7e53e35
Loading
Loading
Loading
Loading
+2 −1
Original line number Original line Diff line number Diff line
@@ -5440,7 +5440,8 @@ S: Supported
F:	drivers/idle/intel_idle.c
F:	drivers/idle/intel_idle.c


INTEL PSTATE DRIVER
INTEL PSTATE DRIVER
M:	Kristen Carlson Accardi <kristen@linux.intel.com>
M:	Srinivas Pandruvada <srinivas.pandruvada@linux.intel.com>
M:	Len Brown <lenb@kernel.org>
L:	linux-pm@vger.kernel.org
L:	linux-pm@vger.kernel.org
S:	Supported
S:	Supported
F:	drivers/cpufreq/intel_pstate.c
F:	drivers/cpufreq/intel_pstate.c
+13 −9
Original line number Original line Diff line number Diff line
@@ -149,6 +149,19 @@ bL_cpufreq_set_rate(u32 cpu, u32 old_cluster, u32 new_cluster, u32 rate)
			__func__, cpu, old_cluster, new_cluster, new_rate);
			__func__, cpu, old_cluster, new_cluster, new_rate);


	ret = clk_set_rate(clk[new_cluster], new_rate * 1000);
	ret = clk_set_rate(clk[new_cluster], new_rate * 1000);
	if (!ret) {
		/*
		 * FIXME: clk_set_rate hasn't returned an error here however it
		 * may be that clk_change_rate failed due to hardware or
		 * firmware issues and wasn't able to report that due to the
		 * current design of the clk core layer. To work around this
		 * problem we will read back the clock rate and check it is
		 * correct. This needs to be removed once clk core is fixed.
		 */
		if (clk_get_rate(clk[new_cluster]) != new_rate * 1000)
			ret = -EIO;
	}

	if (WARN_ON(ret)) {
	if (WARN_ON(ret)) {
		pr_err("clk_set_rate failed: %d, new cluster: %d\n", ret,
		pr_err("clk_set_rate failed: %d, new cluster: %d\n", ret,
				new_cluster);
				new_cluster);
@@ -189,15 +202,6 @@ bL_cpufreq_set_rate(u32 cpu, u32 old_cluster, u32 new_cluster, u32 rate)
		mutex_unlock(&cluster_lock[old_cluster]);
		mutex_unlock(&cluster_lock[old_cluster]);
	}
	}


	/*
	 * FIXME: clk_set_rate has to handle the case where clk_change_rate
	 * can fail due to hardware or firmware issues. Until the clk core
	 * layer is fixed, we can check here. In most of the cases we will
	 * be reading only the cached value anyway. This needs to  be removed
	 * once clk core is fixed.
	 */
	if (bL_cpufreq_get_rate(cpu) != new_rate)
		return -EIO;
	return 0;
	return 0;
}
}


+23 −10
Original line number Original line Diff line number Diff line
@@ -171,10 +171,6 @@ void gov_queue_work(struct dbs_data *dbs_data, struct cpufreq_policy *policy,
{
{
	int i;
	int i;


	mutex_lock(&cpufreq_governor_lock);
	if (!policy->governor_enabled)
		goto out_unlock;

	if (!all_cpus) {
	if (!all_cpus) {
		/*
		/*
		 * Use raw_smp_processor_id() to avoid preemptible warnings.
		 * Use raw_smp_processor_id() to avoid preemptible warnings.
@@ -188,9 +184,6 @@ void gov_queue_work(struct dbs_data *dbs_data, struct cpufreq_policy *policy,
		for_each_cpu(i, policy->cpus)
		for_each_cpu(i, policy->cpus)
			__gov_queue_work(i, dbs_data, delay);
			__gov_queue_work(i, dbs_data, delay);
	}
	}

out_unlock:
	mutex_unlock(&cpufreq_governor_lock);
}
}
EXPORT_SYMBOL_GPL(gov_queue_work);
EXPORT_SYMBOL_GPL(gov_queue_work);


@@ -229,13 +222,24 @@ static void dbs_timer(struct work_struct *work)
	struct cpu_dbs_info *cdbs = container_of(work, struct cpu_dbs_info,
	struct cpu_dbs_info *cdbs = container_of(work, struct cpu_dbs_info,
						 dwork.work);
						 dwork.work);
	struct cpu_common_dbs_info *shared = cdbs->shared;
	struct cpu_common_dbs_info *shared = cdbs->shared;
	struct cpufreq_policy *policy = shared->policy;
	struct cpufreq_policy *policy;
	struct dbs_data *dbs_data = policy->governor_data;
	struct dbs_data *dbs_data;
	unsigned int sampling_rate, delay;
	unsigned int sampling_rate, delay;
	bool modify_all = true;
	bool modify_all = true;


	mutex_lock(&shared->timer_mutex);
	mutex_lock(&shared->timer_mutex);


	policy = shared->policy;

	/*
	 * Governor might already be disabled and there is no point continuing
	 * with the work-handler.
	 */
	if (!policy)
		goto unlock;

	dbs_data = policy->governor_data;

	if (dbs_data->cdata->governor == GOV_CONSERVATIVE) {
	if (dbs_data->cdata->governor == GOV_CONSERVATIVE) {
		struct cs_dbs_tuners *cs_tuners = dbs_data->tuners;
		struct cs_dbs_tuners *cs_tuners = dbs_data->tuners;


@@ -252,6 +256,7 @@ static void dbs_timer(struct work_struct *work)
	delay = dbs_data->cdata->gov_dbs_timer(cdbs, dbs_data, modify_all);
	delay = dbs_data->cdata->gov_dbs_timer(cdbs, dbs_data, modify_all);
	gov_queue_work(dbs_data, policy, delay, modify_all);
	gov_queue_work(dbs_data, policy, delay, modify_all);


unlock:
	mutex_unlock(&shared->timer_mutex);
	mutex_unlock(&shared->timer_mutex);
}
}


@@ -478,9 +483,17 @@ static int cpufreq_governor_stop(struct cpufreq_policy *policy,
	if (!shared || !shared->policy)
	if (!shared || !shared->policy)
		return -EBUSY;
		return -EBUSY;


	/*
	 * Work-handler must see this updated, as it should not proceed any
	 * further after governor is disabled. And so timer_mutex is taken while
	 * updating this value.
	 */
	mutex_lock(&shared->timer_mutex);
	shared->policy = NULL;
	mutex_unlock(&shared->timer_mutex);

	gov_cancel_work(dbs_data, policy);
	gov_cancel_work(dbs_data, policy);


	shared->policy = NULL;
	mutex_destroy(&shared->timer_mutex);
	mutex_destroy(&shared->timer_mutex);
	return 0;
	return 0;
}
}
+6 −4
Original line number Original line Diff line number Diff line
@@ -684,8 +684,6 @@ static void __init intel_pstate_sysfs_expose_params(void)


static void intel_pstate_hwp_enable(struct cpudata *cpudata)
static void intel_pstate_hwp_enable(struct cpudata *cpudata)
{
{
	pr_info("intel_pstate: HWP enabled\n");

	wrmsrl_on_cpu(cpudata->cpu, MSR_PM_ENABLE, 0x1);
	wrmsrl_on_cpu(cpudata->cpu, MSR_PM_ENABLE, 0x1);
}
}


@@ -1557,8 +1555,10 @@ static int __init intel_pstate_init(void)
	if (!all_cpu_data)
	if (!all_cpu_data)
		return -ENOMEM;
		return -ENOMEM;


	if (static_cpu_has_safe(X86_FEATURE_HWP) && !no_hwp)
	if (static_cpu_has_safe(X86_FEATURE_HWP) && !no_hwp) {
		pr_info("intel_pstate: HWP enabled\n");
		hwp_active++;
		hwp_active++;
	}


	if (!hwp_active && hwp_only)
	if (!hwp_active && hwp_only)
		goto out;
		goto out;
@@ -1593,8 +1593,10 @@ static int __init intel_pstate_setup(char *str)


	if (!strcmp(str, "disable"))
	if (!strcmp(str, "disable"))
		no_load = 1;
		no_load = 1;
	if (!strcmp(str, "no_hwp"))
	if (!strcmp(str, "no_hwp")) {
		pr_info("intel_pstate: HWP disabled\n");
		no_hwp = 1;
		no_hwp = 1;
	}
	if (!strcmp(str, "force"))
	if (!strcmp(str, "force"))
		force_load = 1;
		force_load = 1;
	if (!strcmp(str, "hwp_only"))
	if (!strcmp(str, "hwp_only"))
+2 −2
Original line number Original line Diff line number Diff line
@@ -212,11 +212,11 @@ static void s5pv210_set_refresh(enum s5pv210_dmc_port ch, unsigned long freq)
	/* Find current DRAM frequency */
	/* Find current DRAM frequency */
	tmp = s5pv210_dram_conf[ch].freq;
	tmp = s5pv210_dram_conf[ch].freq;


	do_div(tmp, freq);
	tmp /= freq;


	tmp1 = s5pv210_dram_conf[ch].refresh;
	tmp1 = s5pv210_dram_conf[ch].refresh;


	do_div(tmp1, tmp);
	tmp1 /= tmp;


	__raw_writel(tmp1, reg);
	__raw_writel(tmp1, reg);
}
}