Commit 9963d1aa authored by Rusty Russell's avatar Rusty Russell Committed by Dave Jones

[CPUFREQ] clean up speedstep-centrino and reduce cpumask_t usage

Impact: cleanup

1) The #ifdef CONFIG_HOTPLUG_CPU seems unnecessary these days.
2) The loop can simply skip over offline cpus, rather than creating a tmp mask.
3) set_mask is set to either a single cpu or all online cpus in a policy.
   Since it's just used for set_cpus_allowed(), any offline cpus in a policy
   don't matter, so we can just use cpumask_of_cpu() or the policy->cpus.
Signed-off-by: default avatarRusty Russell <rusty@rustcorp.com.au>
Signed-off-by: default avatarMike Travis <travis@sgi.com>
Signed-off-by: default avatarDave Jones <davej@redhat.com>
parent bbeba4c3
...@@ -459,9 +459,7 @@ static int centrino_verify (struct cpufreq_policy *policy) ...@@ -459,9 +459,7 @@ static int centrino_verify (struct cpufreq_policy *policy)
* Sets a new CPUFreq policy. * Sets a new CPUFreq policy.
*/ */
struct allmasks { struct allmasks {
cpumask_t online_policy_cpus;
cpumask_t saved_mask; cpumask_t saved_mask;
cpumask_t set_mask;
cpumask_t covered_cpus; cpumask_t covered_cpus;
}; };
...@@ -475,9 +473,7 @@ static int centrino_target (struct cpufreq_policy *policy, ...@@ -475,9 +473,7 @@ static int centrino_target (struct cpufreq_policy *policy,
int retval = 0; int retval = 0;
unsigned int j, k, first_cpu, tmp; unsigned int j, k, first_cpu, tmp;
CPUMASK_ALLOC(allmasks); CPUMASK_ALLOC(allmasks);
CPUMASK_PTR(online_policy_cpus, allmasks);
CPUMASK_PTR(saved_mask, allmasks); CPUMASK_PTR(saved_mask, allmasks);
CPUMASK_PTR(set_mask, allmasks);
CPUMASK_PTR(covered_cpus, allmasks); CPUMASK_PTR(covered_cpus, allmasks);
if (unlikely(allmasks == NULL)) if (unlikely(allmasks == NULL))
...@@ -497,30 +493,28 @@ static int centrino_target (struct cpufreq_policy *policy, ...@@ -497,30 +493,28 @@ static int centrino_target (struct cpufreq_policy *policy,
goto out; goto out;
} }
#ifdef CONFIG_HOTPLUG_CPU
/* cpufreq holds the hotplug lock, so we are safe from here on */
cpus_and(*online_policy_cpus, cpu_online_map, policy->cpus);
#else
*online_policy_cpus = policy->cpus;
#endif
*saved_mask = current->cpus_allowed; *saved_mask = current->cpus_allowed;
first_cpu = 1; first_cpu = 1;
cpus_clear(*covered_cpus); cpus_clear(*covered_cpus);
for_each_cpu_mask_nr(j, *online_policy_cpus) { for_each_cpu_mask_nr(j, policy->cpus) {
const cpumask_t *mask;
/* cpufreq holds the hotplug lock, so we are safe here */
if (!cpu_online(j))
continue;
/* /*
* Support for SMP systems. * Support for SMP systems.
* Make sure we are running on CPU that wants to change freq * Make sure we are running on CPU that wants to change freq
*/ */
cpus_clear(*set_mask);
if (policy->shared_type == CPUFREQ_SHARED_TYPE_ANY) if (policy->shared_type == CPUFREQ_SHARED_TYPE_ANY)
cpus_or(*set_mask, *set_mask, *online_policy_cpus); mask = &policy->cpus;
else else
cpu_set(j, *set_mask); mask = &cpumask_of_cpu(j);
set_cpus_allowed_ptr(current, set_mask); set_cpus_allowed_ptr(current, mask);
preempt_disable(); preempt_disable();
if (unlikely(!cpu_isset(smp_processor_id(), *set_mask))) { if (unlikely(!cpu_isset(smp_processor_id(), *mask))) {
dprintk("couldn't limit to CPUs in this domain\n"); dprintk("couldn't limit to CPUs in this domain\n");
retval = -EAGAIN; retval = -EAGAIN;
if (first_cpu) { if (first_cpu) {
...@@ -548,7 +542,9 @@ static int centrino_target (struct cpufreq_policy *policy, ...@@ -548,7 +542,9 @@ static int centrino_target (struct cpufreq_policy *policy,
dprintk("target=%dkHz old=%d new=%d msr=%04x\n", dprintk("target=%dkHz old=%d new=%d msr=%04x\n",
target_freq, freqs.old, freqs.new, msr); target_freq, freqs.old, freqs.new, msr);
for_each_cpu_mask_nr(k, *online_policy_cpus) { for_each_cpu_mask_nr(k, policy->cpus) {
if (!cpu_online(k))
continue;
freqs.cpu = k; freqs.cpu = k;
cpufreq_notify_transition(&freqs, cpufreq_notify_transition(&freqs,
CPUFREQ_PRECHANGE); CPUFREQ_PRECHANGE);
...@@ -571,7 +567,9 @@ static int centrino_target (struct cpufreq_policy *policy, ...@@ -571,7 +567,9 @@ static int centrino_target (struct cpufreq_policy *policy,
preempt_enable(); preempt_enable();
} }
for_each_cpu_mask_nr(k, *online_policy_cpus) { for_each_cpu_mask_nr(k, policy->cpus) {
if (!cpu_online(k))
continue;
freqs.cpu = k; freqs.cpu = k;
cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE); cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE);
} }
...@@ -584,18 +582,17 @@ static int centrino_target (struct cpufreq_policy *policy, ...@@ -584,18 +582,17 @@ static int centrino_target (struct cpufreq_policy *policy,
* Best effort undo.. * Best effort undo..
*/ */
if (!cpus_empty(*covered_cpus)) for_each_cpu_mask_nr(j, *covered_cpus) {
for_each_cpu_mask_nr(j, *covered_cpus) { set_cpus_allowed_ptr(current, &cpumask_of_cpu(j));
set_cpus_allowed_ptr(current, wrmsr(MSR_IA32_PERF_CTL, oldmsr, h);
&cpumask_of_cpu(j)); }
wrmsr(MSR_IA32_PERF_CTL, oldmsr, h);
}
tmp = freqs.new; tmp = freqs.new;
freqs.new = freqs.old; freqs.new = freqs.old;
freqs.old = tmp; freqs.old = tmp;
for_each_cpu_mask_nr(j, *online_policy_cpus) { for_each_cpu_mask_nr(j, policy->cpus) {
freqs.cpu = j; if (!cpu_online(j))
continue;
cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE); cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE);
cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE); cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment