Commit 702ed6ef authored by Linus Torvalds's avatar Linus Torvalds

Merge master.kernel.org:/pub/scm/linux/kernel/git/davej/cpufreq

* master.kernel.org:/pub/scm/linux/kernel/git/davej/cpufreq:
  [CPUFREQ] Fix sysfs_create_file return value handling
  [CPUFREQ] ondemand: fix tickless accounting and software coordination bug
  [CPUFREQ] ondemand: add a check to avoid negative load calculation
  [CPUFREQ] Keep userspace governor quiet when it is not being used
  [CPUFREQ] Longhaul - Proper register access
  [CPUFREQ] Kconfig powernow-k8 driver should depend on ACPI P-States driver
  [CPUFREQ] Longhaul - Replace ACPI functions with direct I/O
  [CPUFREQ] Longhaul - Remove duplicate multipliers
  [CPUFREQ] Longhaul - Embedded "conservative"
  [CPUFREQ] acpi-cpufreq: Proper ReadModifyWrite of PERF_CTL MSR
  [CPUFREQ] check return value of sysfs_create_file
  [CPUFREQ] Longhaul - Check ACPI "BM DMA in progress" bit
  [CPUFREQ] Longhaul - Move old_ratio to correct place
  [CPUFREQ] Longhaul - VT8237 support
  [CPUFREQ] Longhaul - Use all kinds of support
  [CPUFREQ] powernow-k8: clarify number of cores.
parents 2f41fc80 58a7295b
...@@ -90,10 +90,17 @@ config X86_POWERNOW_K8 ...@@ -90,10 +90,17 @@ config X86_POWERNOW_K8
If in doubt, say N. If in doubt, say N.
config X86_POWERNOW_K8_ACPI config X86_POWERNOW_K8_ACPI
bool bool "ACPI Support"
depends on X86_POWERNOW_K8 && ACPI_PROCESSOR select ACPI_PROCESSOR
depends on !(X86_POWERNOW_K8 = y && ACPI_PROCESSOR = m) depends on X86_POWERNOW_K8
default y default y
help
This provides access to the K8s Processor Performance States via ACPI.
This driver is probably required for CPUFreq to work with multi-socket and
SMP systems. It is not required on at least some single-socket yet
multi-core systems, even if SMP is enabled.
It is safe to say Y here.
config X86_GX_SUSPMOD config X86_GX_SUSPMOD
tristate "Cyrix MediaGX/NatSemi Geode Suspend Modulation" tristate "Cyrix MediaGX/NatSemi Geode Suspend Modulation"
......
...@@ -167,11 +167,13 @@ static void do_drv_read(struct drv_cmd *cmd) ...@@ -167,11 +167,13 @@ static void do_drv_read(struct drv_cmd *cmd)
static void do_drv_write(struct drv_cmd *cmd) static void do_drv_write(struct drv_cmd *cmd)
{ {
u32 h = 0; u32 lo, hi;
switch (cmd->type) { switch (cmd->type) {
case SYSTEM_INTEL_MSR_CAPABLE: case SYSTEM_INTEL_MSR_CAPABLE:
wrmsr(cmd->addr.msr.reg, cmd->val, h); rdmsr(cmd->addr.msr.reg, lo, hi);
lo = (lo & ~INTEL_MSR_RANGE) | (cmd->val & INTEL_MSR_RANGE);
wrmsr(cmd->addr.msr.reg, lo, hi);
break; break;
case SYSTEM_IO_CAPABLE: case SYSTEM_IO_CAPABLE:
acpi_os_write_port((acpi_io_address)cmd->addr.io.port, acpi_os_write_port((acpi_io_address)cmd->addr.io.port,
...@@ -372,7 +374,6 @@ static int acpi_cpufreq_target(struct cpufreq_policy *policy, ...@@ -372,7 +374,6 @@ static int acpi_cpufreq_target(struct cpufreq_policy *policy,
struct cpufreq_freqs freqs; struct cpufreq_freqs freqs;
cpumask_t online_policy_cpus; cpumask_t online_policy_cpus;
struct drv_cmd cmd; struct drv_cmd cmd;
unsigned int msr;
unsigned int next_state = 0; /* Index into freq_table */ unsigned int next_state = 0; /* Index into freq_table */
unsigned int next_perf_state = 0; /* Index into perf table */ unsigned int next_perf_state = 0; /* Index into perf table */
unsigned int i; unsigned int i;
...@@ -417,11 +418,7 @@ static int acpi_cpufreq_target(struct cpufreq_policy *policy, ...@@ -417,11 +418,7 @@ static int acpi_cpufreq_target(struct cpufreq_policy *policy,
case SYSTEM_INTEL_MSR_CAPABLE: case SYSTEM_INTEL_MSR_CAPABLE:
cmd.type = SYSTEM_INTEL_MSR_CAPABLE; cmd.type = SYSTEM_INTEL_MSR_CAPABLE;
cmd.addr.msr.reg = MSR_IA32_PERF_CTL; cmd.addr.msr.reg = MSR_IA32_PERF_CTL;
msr = cmd.val = (u32) perf->states[next_perf_state].control;
(u32) perf->states[next_perf_state].
control & INTEL_MSR_RANGE;
cmd.val = get_cur_val(online_policy_cpus);
cmd.val = (cmd.val & ~INTEL_MSR_RANGE) | msr;
break; break;
case SYSTEM_IO_CAPABLE: case SYSTEM_IO_CAPABLE:
cmd.type = SYSTEM_IO_CAPABLE; cmd.type = SYSTEM_IO_CAPABLE;
......
...@@ -29,6 +29,7 @@ ...@@ -29,6 +29,7 @@
#include <linux/pci.h> #include <linux/pci.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <linux/string.h> #include <linux/string.h>
#include <linux/delay.h>
#include <asm/msr.h> #include <asm/msr.h>
#include <asm/timex.h> #include <asm/timex.h>
...@@ -55,7 +56,6 @@ ...@@ -55,7 +56,6 @@
/* Flags */ /* Flags */
#define USE_ACPI_C3 (1 << 1) #define USE_ACPI_C3 (1 << 1)
#define USE_NORTHBRIDGE (1 << 2) #define USE_NORTHBRIDGE (1 << 2)
#define USE_VT8235 (1 << 3)
static int cpu_model; static int cpu_model;
static unsigned int numscales=16; static unsigned int numscales=16;
...@@ -63,19 +63,15 @@ static unsigned int fsb; ...@@ -63,19 +63,15 @@ static unsigned int fsb;
static const struct mV_pos *vrm_mV_table; static const struct mV_pos *vrm_mV_table;
static const unsigned char *mV_vrm_table; static const unsigned char *mV_vrm_table;
struct f_msr {
u8 vrm;
u8 pos;
};
static struct f_msr f_msr_table[32];
static unsigned int highest_speed, lowest_speed; /* kHz */ static unsigned int highest_speed, lowest_speed; /* kHz */
static unsigned int minmult, maxmult; static unsigned int minmult, maxmult;
static int can_scale_voltage; static int can_scale_voltage;
static struct acpi_processor *pr = NULL; static struct acpi_processor *pr = NULL;
static struct acpi_processor_cx *cx = NULL; static struct acpi_processor_cx *cx = NULL;
static u32 acpi_regs_addr;
static u8 longhaul_flags; static u8 longhaul_flags;
static u8 longhaul_pos; static unsigned int longhaul_index;
/* Module parameters */ /* Module parameters */
static int scale_voltage; static int scale_voltage;
...@@ -144,7 +140,7 @@ static void do_longhaul1(unsigned int clock_ratio_index) ...@@ -144,7 +140,7 @@ static void do_longhaul1(unsigned int clock_ratio_index)
rdmsrl(MSR_VIA_BCR2, bcr2.val); rdmsrl(MSR_VIA_BCR2, bcr2.val);
/* Enable software clock multiplier */ /* Enable software clock multiplier */
bcr2.bits.ESOFTBF = 1; bcr2.bits.ESOFTBF = 1;
bcr2.bits.CLOCKMUL = clock_ratio_index; bcr2.bits.CLOCKMUL = clock_ratio_index & 0xff;
/* Sync to timer tick */ /* Sync to timer tick */
safe_halt(); safe_halt();
...@@ -163,14 +159,12 @@ static void do_longhaul1(unsigned int clock_ratio_index) ...@@ -163,14 +159,12 @@ static void do_longhaul1(unsigned int clock_ratio_index)
/* For processor with Longhaul MSR */ /* For processor with Longhaul MSR */
static void do_powersaver(int cx_address, unsigned int clock_ratio_index) static void do_powersaver(int cx_address, unsigned int clock_ratio_index,
unsigned int dir)
{ {
union msr_longhaul longhaul; union msr_longhaul longhaul;
u8 dest_pos;
u32 t; u32 t;
dest_pos = f_msr_table[clock_ratio_index].pos;
rdmsrl(MSR_VIA_LONGHAUL, longhaul.val); rdmsrl(MSR_VIA_LONGHAUL, longhaul.val);
/* Setup new frequency */ /* Setup new frequency */
longhaul.bits.RevisionKey = longhaul.bits.RevisionID; longhaul.bits.RevisionKey = longhaul.bits.RevisionID;
...@@ -178,11 +172,11 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index) ...@@ -178,11 +172,11 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index)
longhaul.bits.SoftBusRatio4 = (clock_ratio_index & 0x10) >> 4; longhaul.bits.SoftBusRatio4 = (clock_ratio_index & 0x10) >> 4;
/* Setup new voltage */ /* Setup new voltage */
if (can_scale_voltage) if (can_scale_voltage)
longhaul.bits.SoftVID = f_msr_table[clock_ratio_index].vrm; longhaul.bits.SoftVID = (clock_ratio_index >> 8) & 0x1f;
/* Sync to timer tick */ /* Sync to timer tick */
safe_halt(); safe_halt();
/* Raise voltage if necessary */ /* Raise voltage if necessary */
if (can_scale_voltage && longhaul_pos < dest_pos) { if (can_scale_voltage && dir) {
longhaul.bits.EnableSoftVID = 1; longhaul.bits.EnableSoftVID = 1;
wrmsrl(MSR_VIA_LONGHAUL, longhaul.val); wrmsrl(MSR_VIA_LONGHAUL, longhaul.val);
/* Change voltage */ /* Change voltage */
...@@ -199,7 +193,6 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index) ...@@ -199,7 +193,6 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index)
} }
longhaul.bits.EnableSoftVID = 0; longhaul.bits.EnableSoftVID = 0;
wrmsrl(MSR_VIA_LONGHAUL, longhaul.val); wrmsrl(MSR_VIA_LONGHAUL, longhaul.val);
longhaul_pos = dest_pos;
} }
/* Change frequency on next halt or sleep */ /* Change frequency on next halt or sleep */
...@@ -220,7 +213,7 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index) ...@@ -220,7 +213,7 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index)
wrmsrl(MSR_VIA_LONGHAUL, longhaul.val); wrmsrl(MSR_VIA_LONGHAUL, longhaul.val);
/* Reduce voltage if necessary */ /* Reduce voltage if necessary */
if (can_scale_voltage && longhaul_pos > dest_pos) { if (can_scale_voltage && !dir) {
longhaul.bits.EnableSoftVID = 1; longhaul.bits.EnableSoftVID = 1;
wrmsrl(MSR_VIA_LONGHAUL, longhaul.val); wrmsrl(MSR_VIA_LONGHAUL, longhaul.val);
/* Change voltage */ /* Change voltage */
...@@ -237,7 +230,6 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index) ...@@ -237,7 +230,6 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index)
} }
longhaul.bits.EnableSoftVID = 0; longhaul.bits.EnableSoftVID = 0;
wrmsrl(MSR_VIA_LONGHAUL, longhaul.val); wrmsrl(MSR_VIA_LONGHAUL, longhaul.val);
longhaul_pos = dest_pos;
} }
} }
...@@ -248,25 +240,28 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index) ...@@ -248,25 +240,28 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index)
* Sets a new clock ratio. * Sets a new clock ratio.
*/ */
static void longhaul_setstate(unsigned int clock_ratio_index) static void longhaul_setstate(unsigned int table_index)
{ {
unsigned int clock_ratio_index;
int speed, mult; int speed, mult;
struct cpufreq_freqs freqs; struct cpufreq_freqs freqs;
static unsigned int old_ratio=-1;
unsigned long flags; unsigned long flags;
unsigned int pic1_mask, pic2_mask; unsigned int pic1_mask, pic2_mask;
u16 bm_status = 0;
u32 bm_timeout = 1000;
unsigned int dir = 0;
if (old_ratio == clock_ratio_index) clock_ratio_index = longhaul_table[table_index].index;
return; /* Safety precautions */
old_ratio = clock_ratio_index; mult = clock_ratio[clock_ratio_index & 0x1f];
mult = clock_ratio[clock_ratio_index];
if (mult == -1) if (mult == -1)
return; return;
speed = calc_speed(mult); speed = calc_speed(mult);
if ((speed > highest_speed) || (speed < lowest_speed)) if ((speed > highest_speed) || (speed < lowest_speed))
return; return;
/* Voltage transition before frequency transition? */
if (can_scale_voltage && longhaul_index < table_index)
dir = 1;
freqs.old = calc_speed(longhaul_get_cpu_mult()); freqs.old = calc_speed(longhaul_get_cpu_mult());
freqs.new = speed; freqs.new = speed;
...@@ -285,11 +280,24 @@ static void longhaul_setstate(unsigned int clock_ratio_index) ...@@ -285,11 +280,24 @@ static void longhaul_setstate(unsigned int clock_ratio_index)
outb(0xFF,0xA1); /* Overkill */ outb(0xFF,0xA1); /* Overkill */
outb(0xFE,0x21); /* TMR0 only */ outb(0xFE,0x21); /* TMR0 only */
/* Wait while PCI bus is busy. */
if (acpi_regs_addr && (longhaul_flags & USE_NORTHBRIDGE
|| ((pr != NULL) && pr->flags.bm_control))) {
bm_status = inw(acpi_regs_addr);
bm_status &= 1 << 4;
while (bm_status && bm_timeout) {
outw(1 << 4, acpi_regs_addr);
bm_timeout--;
bm_status = inw(acpi_regs_addr);
bm_status &= 1 << 4;
}
}
if (longhaul_flags & USE_NORTHBRIDGE) { if (longhaul_flags & USE_NORTHBRIDGE) {
/* Disable AGP and PCI arbiters */ /* Disable AGP and PCI arbiters */
outb(3, 0x22); outb(3, 0x22);
} else if ((pr != NULL) && pr->flags.bm_control) { } else if ((pr != NULL) && pr->flags.bm_control) {
/* Disable bus master arbitration */ /* Disable bus master arbitration */
acpi_set_register(ACPI_BITREG_ARB_DISABLE, 1); acpi_set_register(ACPI_BITREG_ARB_DISABLE, 1);
} }
switch (longhaul_version) { switch (longhaul_version) {
...@@ -314,9 +322,9 @@ static void longhaul_setstate(unsigned int clock_ratio_index) ...@@ -314,9 +322,9 @@ static void longhaul_setstate(unsigned int clock_ratio_index)
if (longhaul_flags & USE_ACPI_C3) { if (longhaul_flags & USE_ACPI_C3) {
/* Don't allow wakeup */ /* Don't allow wakeup */
acpi_set_register(ACPI_BITREG_BUS_MASTER_RLD, 0); acpi_set_register(ACPI_BITREG_BUS_MASTER_RLD, 0);
do_powersaver(cx->address, clock_ratio_index); do_powersaver(cx->address, clock_ratio_index, dir);
} else { } else {
do_powersaver(0, clock_ratio_index); do_powersaver(0, clock_ratio_index, dir);
} }
break; break;
} }
...@@ -336,6 +344,9 @@ static void longhaul_setstate(unsigned int clock_ratio_index) ...@@ -336,6 +344,9 @@ static void longhaul_setstate(unsigned int clock_ratio_index)
freqs.new = calc_speed(longhaul_get_cpu_mult()); freqs.new = calc_speed(longhaul_get_cpu_mult());
cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE); cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE);
if (!bm_timeout)
printk(KERN_INFO PFX "Warning: Timeout while waiting for idle PCI bus.\n");
} }
/* /*
...@@ -369,7 +380,8 @@ static int guess_fsb(int mult) ...@@ -369,7 +380,8 @@ static int guess_fsb(int mult)
static int __init longhaul_get_ranges(void) static int __init longhaul_get_ranges(void)
{ {
unsigned int j, k = 0; unsigned int i, j, k = 0;
unsigned int ratio;
int mult; int mult;
/* Get current frequency */ /* Get current frequency */
...@@ -423,8 +435,7 @@ static int __init longhaul_get_ranges(void) ...@@ -423,8 +435,7 @@ static int __init longhaul_get_ranges(void)
if(!longhaul_table) if(!longhaul_table)
return -ENOMEM; return -ENOMEM;
for (j=0; j < numscales; j++) { for (j = 0; j < numscales; j++) {
unsigned int ratio;
ratio = clock_ratio[j]; ratio = clock_ratio[j];
if (ratio == -1) if (ratio == -1)
continue; continue;
...@@ -434,13 +445,41 @@ static int __init longhaul_get_ranges(void) ...@@ -434,13 +445,41 @@ static int __init longhaul_get_ranges(void)
longhaul_table[k].index = j; longhaul_table[k].index = j;
k++; k++;
} }
if (k <= 1) {
kfree(longhaul_table);
return -ENODEV;
}
/* Sort */
for (j = 0; j < k - 1; j++) {
unsigned int min_f, min_i;
min_f = longhaul_table[j].frequency;
min_i = j;
for (i = j + 1; i < k; i++) {
if (longhaul_table[i].frequency < min_f) {
min_f = longhaul_table[i].frequency;
min_i = i;
}
}
if (min_i != j) {
unsigned int temp;
temp = longhaul_table[j].frequency;
longhaul_table[j].frequency = longhaul_table[min_i].frequency;
longhaul_table[min_i].frequency = temp;
temp = longhaul_table[j].index;
longhaul_table[j].index = longhaul_table[min_i].index;
longhaul_table[min_i].index = temp;
}
}
longhaul_table[k].frequency = CPUFREQ_TABLE_END; longhaul_table[k].frequency = CPUFREQ_TABLE_END;
if (!k) {
kfree (longhaul_table);
return -EINVAL;
}
/* Find index we are running on */
for (j = 0; j < k; j++) {
if (clock_ratio[longhaul_table[j].index & 0x1f] == mult) {
longhaul_index = j;
break;
}
}
return 0; return 0;
} }
...@@ -448,7 +487,7 @@ static int __init longhaul_get_ranges(void) ...@@ -448,7 +487,7 @@ static int __init longhaul_get_ranges(void)
static void __init longhaul_setup_voltagescaling(void) static void __init longhaul_setup_voltagescaling(void)
{ {
union msr_longhaul longhaul; union msr_longhaul longhaul;
struct mV_pos minvid, maxvid; struct mV_pos minvid, maxvid, vid;
unsigned int j, speed, pos, kHz_step, numvscales; unsigned int j, speed, pos, kHz_step, numvscales;
int min_vid_speed; int min_vid_speed;
...@@ -459,11 +498,11 @@ static void __init longhaul_setup_voltagescaling(void) ...@@ -459,11 +498,11 @@ static void __init longhaul_setup_voltagescaling(void)
} }
if (!longhaul.bits.VRMRev) { if (!longhaul.bits.VRMRev) {
printk (KERN_INFO PFX "VRM 8.5\n"); printk(KERN_INFO PFX "VRM 8.5\n");
vrm_mV_table = &vrm85_mV[0]; vrm_mV_table = &vrm85_mV[0];
mV_vrm_table = &mV_vrm85[0]; mV_vrm_table = &mV_vrm85[0];
} else { } else {
printk (KERN_INFO PFX "Mobile VRM\n"); printk(KERN_INFO PFX "Mobile VRM\n");
if (cpu_model < CPU_NEHEMIAH) if (cpu_model < CPU_NEHEMIAH)
return; return;
vrm_mV_table = &mobilevrm_mV[0]; vrm_mV_table = &mobilevrm_mV[0];
...@@ -523,7 +562,6 @@ static void __init longhaul_setup_voltagescaling(void) ...@@ -523,7 +562,6 @@ static void __init longhaul_setup_voltagescaling(void)
/* Calculate kHz for one voltage step */ /* Calculate kHz for one voltage step */
kHz_step = (highest_speed - min_vid_speed) / numvscales; kHz_step = (highest_speed - min_vid_speed) / numvscales;
j = 0; j = 0;
while (longhaul_table[j].frequency != CPUFREQ_TABLE_END) { while (longhaul_table[j].frequency != CPUFREQ_TABLE_END) {
speed = longhaul_table[j].frequency; speed = longhaul_table[j].frequency;
...@@ -531,15 +569,14 @@ static void __init longhaul_setup_voltagescaling(void) ...@@ -531,15 +569,14 @@ static void __init longhaul_setup_voltagescaling(void)
pos = (speed - min_vid_speed) / kHz_step + minvid.pos; pos = (speed - min_vid_speed) / kHz_step + minvid.pos;
else else
pos = minvid.pos; pos = minvid.pos;
f_msr_table[longhaul_table[j].index].vrm = mV_vrm_table[pos]; longhaul_table[j].index |= mV_vrm_table[pos] << 8;
f_msr_table[longhaul_table[j].index].pos = pos; vid = vrm_mV_table[mV_vrm_table[pos]];
printk(KERN_INFO PFX "f: %d kHz, index: %d, vid: %d mV\n", speed, j, vid.mV);
j++; j++;
} }
longhaul_pos = maxvid.pos;
can_scale_voltage = 1; can_scale_voltage = 1;
printk(KERN_INFO PFX "Voltage scaling enabled. " printk(KERN_INFO PFX "Voltage scaling enabled.\n");
"Use of \"conservative\" governor is highly recommended.\n");
} }
...@@ -553,15 +590,44 @@ static int longhaul_target(struct cpufreq_policy *policy, ...@@ -553,15 +590,44 @@ static int longhaul_target(struct cpufreq_policy *policy,
unsigned int target_freq, unsigned int relation) unsigned int target_freq, unsigned int relation)
{ {
unsigned int table_index = 0; unsigned int table_index = 0;
unsigned int new_clock_ratio = 0; unsigned int i;
unsigned int dir = 0;
u8 vid, current_vid;
if (cpufreq_frequency_table_target(policy, longhaul_table, target_freq, relation, &table_index)) if (cpufreq_frequency_table_target(policy, longhaul_table, target_freq, relation, &table_index))
return -EINVAL; return -EINVAL;
new_clock_ratio = longhaul_table[table_index].index & 0xFF; /* Don't set same frequency again */
if (longhaul_index == table_index)
longhaul_setstate(new_clock_ratio); return 0;
if (!can_scale_voltage)
longhaul_setstate(table_index);
else {
/* On test system voltage transitions exceeding single
* step up or down were turning motherboard off. Both
* "ondemand" and "userspace" are unsafe. C7 is doing
* this in hardware, C3 is old and we need to do this
* in software. */
i = longhaul_index;
current_vid = (longhaul_table[longhaul_index].index >> 8) & 0x1f;
if (table_index > longhaul_index)
dir = 1;
while (i != table_index) {
vid = (longhaul_table[i].index >> 8) & 0x1f;
if (vid != current_vid) {
longhaul_setstate(i);
current_vid = vid;
msleep(200);
}
if (dir)
i++;
else
i--;
}
longhaul_setstate(table_index);
}
longhaul_index = table_index;
return 0; return 0;
} }
...@@ -590,11 +656,10 @@ static acpi_status longhaul_walk_callback(acpi_handle obj_handle, ...@@ -590,11 +656,10 @@ static acpi_status longhaul_walk_callback(acpi_handle obj_handle,
static int enable_arbiter_disable(void) static int enable_arbiter_disable(void)
{ {
struct pci_dev *dev; struct pci_dev *dev;
int status; int status = 1;
int reg; int reg;
u8 pci_cmd; u8 pci_cmd;
status = 1;
/* Find PLE133 host bridge */ /* Find PLE133 host bridge */
reg = 0x78; reg = 0x78;
dev = pci_get_device(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8601_0, dev = pci_get_device(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8601_0,
...@@ -627,13 +692,17 @@ static int enable_arbiter_disable(void) ...@@ -627,13 +692,17 @@ static int enable_arbiter_disable(void)
return 0; return 0;
} }
static int longhaul_setup_vt8235(void) static int longhaul_setup_southbridge(void)
{ {
struct pci_dev *dev; struct pci_dev *dev;
u8 pci_cmd; u8 pci_cmd;
/* Find VT8235 southbridge */ /* Find VT8235 southbridge */
dev = pci_get_device(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235, NULL); dev = pci_get_device(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235, NULL);
if (dev == NULL)
/* Find VT8237 southbridge */
dev = pci_get_device(PCI_VENDOR_ID_VIA,
PCI_DEVICE_ID_VIA_8237, NULL);
if (dev != NULL) { if (dev != NULL) {
/* Set transition time to max */ /* Set transition time to max */
pci_read_config_byte(dev, 0xec, &pci_cmd); pci_read_config_byte(dev, 0xec, &pci_cmd);
...@@ -645,6 +714,14 @@ static int longhaul_setup_vt8235(void) ...@@ -645,6 +714,14 @@ static int longhaul_setup_vt8235(void)
pci_read_config_byte(dev, 0xe5, &pci_cmd); pci_read_config_byte(dev, 0xe5, &pci_cmd);
pci_cmd |= 1 << 7; pci_cmd |= 1 << 7;
pci_write_config_byte(dev, 0xe5, pci_cmd); pci_write_config_byte(dev, 0xe5, pci_cmd);
/* Get address of ACPI registers block*/
pci_read_config_byte(dev, 0x81, &pci_cmd);
if (pci_cmd & 1 << 7) {
pci_read_config_dword(dev, 0x88, &acpi_regs_addr);
acpi_regs_addr &= 0xff00;
printk(KERN_INFO PFX "ACPI I/O at 0x%x\n", acpi_regs_addr);
}
pci_dev_put(dev); pci_dev_put(dev);
return 1; return 1;
} }
...@@ -657,7 +734,6 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy) ...@@ -657,7 +734,6 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy)
char *cpuname=NULL; char *cpuname=NULL;
int ret; int ret;
u32 lo, hi; u32 lo, hi;
int vt8235_present;
/* Check what we have on this motherboard */ /* Check what we have on this motherboard */
switch (c->x86_model) { switch (c->x86_model) {
...@@ -755,7 +831,7 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy) ...@@ -755,7 +831,7 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy)
}; };
/* Doesn't hurt */ /* Doesn't hurt */
vt8235_present = longhaul_setup_vt8235(); longhaul_setup_southbridge();
/* Find ACPI data for processor */ /* Find ACPI data for processor */
acpi_walk_namespace(ACPI_TYPE_PROCESSOR, ACPI_ROOT_OBJECT, acpi_walk_namespace(ACPI_TYPE_PROCESSOR, ACPI_ROOT_OBJECT,
...@@ -765,35 +841,26 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy) ...@@ -765,35 +841,26 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy)
/* Check ACPI support for C3 state */ /* Check ACPI support for C3 state */
if (pr != NULL && longhaul_version == TYPE_POWERSAVER) { if (pr != NULL && longhaul_version == TYPE_POWERSAVER) {
cx = &pr->power.states[ACPI_STATE_C3]; cx = &pr->power.states[ACPI_STATE_C3];
if (cx->address > 0 && cx->latency <= 1000) { if (cx->address > 0 && cx->latency <= 1000)
longhaul_flags |= USE_ACPI_C3; longhaul_flags |= USE_ACPI_C3;
goto print_support_type;
}
} }
/* Check if northbridge is friendly */ /* Check if northbridge is friendly */
if (enable_arbiter_disable()) { if (enable_arbiter_disable())
longhaul_flags |= USE_NORTHBRIDGE; longhaul_flags |= USE_NORTHBRIDGE;
goto print_support_type;
}
/* Use VT8235 southbridge if present */
if (longhaul_version == TYPE_POWERSAVER && vt8235_present) {
longhaul_flags |= USE_VT8235;
goto print_support_type;
}
/* Check ACPI support for bus master arbiter disable */ /* Check ACPI support for bus master arbiter disable */
if ((pr == NULL) || !(pr->flags.bm_control)) { if (!(longhaul_flags & USE_ACPI_C3
|| longhaul_flags & USE_NORTHBRIDGE)
&& ((pr == NULL) || !(pr->flags.bm_control))) {
printk(KERN_ERR PFX printk(KERN_ERR PFX
"No ACPI support. Unsupported northbridge.\n"); "No ACPI support. Unsupported northbridge.\n");
return -ENODEV; return -ENODEV;
} }
print_support_type:
if (longhaul_flags & USE_NORTHBRIDGE) if (longhaul_flags & USE_NORTHBRIDGE)
printk (KERN_INFO PFX "Using northbridge support.\n"); printk(KERN_INFO PFX "Using northbridge support.\n");
else if (longhaul_flags & USE_VT8235) if (longhaul_flags & USE_ACPI_C3)
printk (KERN_INFO PFX "Using VT8235 support.\n"); printk(KERN_INFO PFX "Using ACPI support.\n");
else
printk (KERN_INFO PFX "Using ACPI support.\n");
ret = longhaul_get_ranges(); ret = longhaul_get_ranges();
if (ret != 0) if (ret != 0)
......
...@@ -180,7 +180,7 @@ static const int __initdata ezrat_clock_ratio[32] = { ...@@ -180,7 +180,7 @@ static const int __initdata ezrat_clock_ratio[32] = {
-1, /* 0000 -> RESERVED (10.0x) */ -1, /* 0000 -> RESERVED (10.0x) */
110, /* 0001 -> 11.0x */ 110, /* 0001 -> 11.0x */
120, /* 0010 -> 12.0x */ -1, /* 0010 -> 12.0x */
-1, /* 0011 -> RESERVED (9.0x)*/ -1, /* 0011 -> RESERVED (9.0x)*/
105, /* 0100 -> 10.5x */ 105, /* 0100 -> 10.5x */
115, /* 0101 -> 11.5x */ 115, /* 0101 -> 11.5x */
...@@ -237,7 +237,7 @@ static const int __initdata ezrat_eblcr[32] = { ...@@ -237,7 +237,7 @@ static const int __initdata ezrat_eblcr[32] = {
static const int __initdata nehemiah_clock_ratio[32] = { static const int __initdata nehemiah_clock_ratio[32] = {
100, /* 0000 -> 10.0x */ 100, /* 0000 -> 10.0x */
160, /* 0001 -> 16.0x */ -1, /* 0001 -> 16.0x */
40, /* 0010 -> 4.0x */ 40, /* 0010 -> 4.0x */
90, /* 0011 -> 9.0x */ 90, /* 0011 -> 9.0x */
95, /* 0100 -> 9.5x */ 95, /* 0100 -> 9.5x */
...@@ -252,10 +252,10 @@ static const int __initdata nehemiah_clock_ratio[32] = { ...@@ -252,10 +252,10 @@ static const int __initdata nehemiah_clock_ratio[32] = {
75, /* 1101 -> 7.5x */ 75, /* 1101 -> 7.5x */
85, /* 1110 -> 8.5x */ 85, /* 1110 -> 8.5x */
120, /* 1111 -> 12.0x */ 120, /* 1111 -> 12.0x */
100, /* 0000 -> 10.0x */ -1, /* 0000 -> 10.0x */
110, /* 0001 -> 11.0x */ 110, /* 0001 -> 11.0x */
120, /* 0010 -> 12.0x */ -1, /* 0010 -> 12.0x */
90, /* 0011 -> 9.0x */ -1, /* 0011 -> 9.0x */
105, /* 0100 -> 10.5x */ 105, /* 0100 -> 10.5x */
115, /* 0101 -> 11.5x */ 115, /* 0101 -> 11.5x */
125, /* 0110 -> 12.5x */ 125, /* 0110 -> 12.5x */
...@@ -267,7 +267,7 @@ static const int __initdata nehemiah_clock_ratio[32] = { ...@@ -267,7 +267,7 @@ static const int __initdata nehemiah_clock_ratio[32] = {
145, /* 1100 -> 14.5x */ 145, /* 1100 -> 14.5x */
155, /* 1101 -> 15.5x */ 155, /* 1101 -> 15.5x */
-1, /* 1110 -> RESERVED (13.0x) */ -1, /* 1110 -> RESERVED (13.0x) */
120, /* 1111 -> 12.0x */ -1, /* 1111 -> 12.0x */
}; };
static const int __initdata nehemiah_eblcr[32] = { static const int __initdata nehemiah_eblcr[32] = {
......
...@@ -1330,8 +1330,9 @@ static int __cpuinit powernowk8_init(void) ...@@ -1330,8 +1330,9 @@ static int __cpuinit powernowk8_init(void)
if (supported_cpus == num_online_cpus()) { if (supported_cpus == num_online_cpus()) {
printk(KERN_INFO PFX "Found %d %s " printk(KERN_INFO PFX "Found %d %s "
"processors (" VERSION ")\n", supported_cpus, "processors (%d cpu cores) (" VERSION ")\n",
boot_cpu_data.x86_model_id); supported_cpus/cpu_data[0].booted_cores,
boot_cpu_data.x86_model_id, supported_cpus);
return cpufreq_register_driver(&cpufreq_amd64_driver); return cpufreq_register_driver(&cpufreq_amd64_driver);
} }
......
...@@ -826,13 +826,21 @@ static int cpufreq_add_dev (struct sys_device * sys_dev) ...@@ -826,13 +826,21 @@ static int cpufreq_add_dev (struct sys_device * sys_dev)
/* set up files for this cpu device */ /* set up files for this cpu device */
drv_attr = cpufreq_driver->attr; drv_attr = cpufreq_driver->attr;
while ((drv_attr) && (*drv_attr)) { while ((drv_attr) && (*drv_attr)) {
sysfs_create_file(&policy->kobj, &((*drv_attr)->attr)); ret = sysfs_create_file(&policy->kobj, &((*drv_attr)->attr));
if (ret)
goto err_out_driver_exit;
drv_attr++; drv_attr++;
} }
if (cpufreq_driver->get) if (cpufreq_driver->get){
sysfs_create_file(&policy->kobj, &cpuinfo_cur_freq.attr); ret = sysfs_create_file(&policy->kobj, &cpuinfo_cur_freq.attr);
if (cpufreq_driver->target) if (ret)
sysfs_create_file(&policy->kobj, &scaling_cur_freq.attr); goto err_out_driver_exit;
}
if (cpufreq_driver->target){
ret = sysfs_create_file(&policy->kobj, &scaling_cur_freq.attr);
if (ret)
goto err_out_driver_exit;
}
spin_lock_irqsave(&cpufreq_driver_lock, flags); spin_lock_irqsave(&cpufreq_driver_lock, flags);
for_each_cpu_mask(j, policy->cpus) { for_each_cpu_mask(j, policy->cpus) {
......
...@@ -96,15 +96,25 @@ static struct dbs_tuners { ...@@ -96,15 +96,25 @@ static struct dbs_tuners {
static inline cputime64_t get_cpu_idle_time(unsigned int cpu) static inline cputime64_t get_cpu_idle_time(unsigned int cpu)
{ {
cputime64_t retval; cputime64_t idle_time;
cputime64_t cur_jiffies;
cputime64_t busy_time;
retval = cputime64_add(kstat_cpu(cpu).cpustat.idle, cur_jiffies = jiffies64_to_cputime64(get_jiffies_64());
kstat_cpu(cpu).cpustat.iowait); busy_time = cputime64_add(kstat_cpu(cpu).cpustat.user,
kstat_cpu(cpu).cpustat.system);
if (dbs_tuners_ins.ignore_nice) busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.irq);
retval = cputime64_add(retval, kstat_cpu(cpu).cpustat.nice); busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.softirq);
busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.steal);
return retval; if (!dbs_tuners_ins.ignore_nice) {
busy_time = cputime64_add(busy_time,
kstat_cpu(cpu).cpustat.nice);
}
idle_time = cputime64_sub(cur_jiffies, busy_time);
return idle_time;
} }
/* /*
...@@ -325,7 +335,7 @@ static struct attribute_group dbs_attr_group = { ...@@ -325,7 +335,7 @@ static struct attribute_group dbs_attr_group = {
static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info) static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info)
{ {
unsigned int idle_ticks, total_ticks; unsigned int idle_ticks, total_ticks;
unsigned int load; unsigned int load = 0;
cputime64_t cur_jiffies; cputime64_t cur_jiffies;
struct cpufreq_policy *policy; struct cpufreq_policy *policy;
...@@ -339,7 +349,8 @@ static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info) ...@@ -339,7 +349,8 @@ static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info)
cur_jiffies = jiffies64_to_cputime64(get_jiffies_64()); cur_jiffies = jiffies64_to_cputime64(get_jiffies_64());
total_ticks = (unsigned int) cputime64_sub(cur_jiffies, total_ticks = (unsigned int) cputime64_sub(cur_jiffies,
this_dbs_info->prev_cpu_wall); this_dbs_info->prev_cpu_wall);
this_dbs_info->prev_cpu_wall = cur_jiffies; this_dbs_info->prev_cpu_wall = get_jiffies_64();
if (!total_ticks) if (!total_ticks)
return; return;
/* /*
...@@ -370,7 +381,8 @@ static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info) ...@@ -370,7 +381,8 @@ static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info)
if (tmp_idle_ticks < idle_ticks) if (tmp_idle_ticks < idle_ticks)
idle_ticks = tmp_idle_ticks; idle_ticks = tmp_idle_ticks;
} }
load = (100 * (total_ticks - idle_ticks)) / total_ticks; if (likely(total_ticks > idle_ticks))
load = (100 * (total_ticks - idle_ticks)) / total_ticks;
/* Check for frequency increase */ /* Check for frequency increase */
if (load > dbs_tuners_ins.up_threshold) { if (load > dbs_tuners_ins.up_threshold) {
......
...@@ -37,6 +37,7 @@ static unsigned int cpu_set_freq[NR_CPUS]; /* CPU freq desired by userspace */ ...@@ -37,6 +37,7 @@ static unsigned int cpu_set_freq[NR_CPUS]; /* CPU freq desired by userspace */
static unsigned int cpu_is_managed[NR_CPUS]; static unsigned int cpu_is_managed[NR_CPUS];
static DEFINE_MUTEX (userspace_mutex); static DEFINE_MUTEX (userspace_mutex);
static int cpus_using_userspace_governor;
#define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_GOVERNOR, "userspace", msg) #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_GOVERNOR, "userspace", msg)
...@@ -47,7 +48,11 @@ userspace_cpufreq_notifier(struct notifier_block *nb, unsigned long val, ...@@ -47,7 +48,11 @@ userspace_cpufreq_notifier(struct notifier_block *nb, unsigned long val,
{ {
struct cpufreq_freqs *freq = data; struct cpufreq_freqs *freq = data;
dprintk("saving cpu_cur_freq of cpu %u to be %u kHz\n", freq->cpu, freq->new); if (!cpu_is_managed[freq->cpu])
return 0;
dprintk("saving cpu_cur_freq of cpu %u to be %u kHz\n",
freq->cpu, freq->new);
cpu_cur_freq[freq->cpu] = freq->new; cpu_cur_freq[freq->cpu] = freq->new;
return 0; return 0;
...@@ -142,6 +147,13 @@ static int cpufreq_governor_userspace(struct cpufreq_policy *policy, ...@@ -142,6 +147,13 @@ static int cpufreq_governor_userspace(struct cpufreq_policy *policy,
if (rc) if (rc)
goto start_out; goto start_out;
if (cpus_using_userspace_governor == 0) {
cpufreq_register_notifier(
&userspace_cpufreq_notifier_block,
CPUFREQ_TRANSITION_NOTIFIER);
}
cpus_using_userspace_governor++;
cpu_is_managed[cpu] = 1; cpu_is_managed[cpu] = 1;
cpu_min_freq[cpu] = policy->min; cpu_min_freq[cpu] = policy->min;
cpu_max_freq[cpu] = policy->max; cpu_max_freq[cpu] = policy->max;
...@@ -153,6 +165,13 @@ static int cpufreq_governor_userspace(struct cpufreq_policy *policy, ...@@ -153,6 +165,13 @@ static int cpufreq_governor_userspace(struct cpufreq_policy *policy,
break; break;
case CPUFREQ_GOV_STOP: case CPUFREQ_GOV_STOP:
mutex_lock(&userspace_mutex); mutex_lock(&userspace_mutex);
cpus_using_userspace_governor--;
if (cpus_using_userspace_governor == 0) {
cpufreq_unregister_notifier(
&userspace_cpufreq_notifier_block,
CPUFREQ_TRANSITION_NOTIFIER);
}
cpu_is_managed[cpu] = 0; cpu_is_managed[cpu] = 0;
cpu_min_freq[cpu] = 0; cpu_min_freq[cpu] = 0;
cpu_max_freq[cpu] = 0; cpu_max_freq[cpu] = 0;
...@@ -198,7 +217,6 @@ EXPORT_SYMBOL(cpufreq_gov_userspace); ...@@ -198,7 +217,6 @@ EXPORT_SYMBOL(cpufreq_gov_userspace);
static int __init cpufreq_gov_userspace_init(void) static int __init cpufreq_gov_userspace_init(void)
{ {
cpufreq_register_notifier(&userspace_cpufreq_notifier_block, CPUFREQ_TRANSITION_NOTIFIER);
return cpufreq_register_governor(&cpufreq_gov_userspace); return cpufreq_register_governor(&cpufreq_gov_userspace);
} }
...@@ -206,7 +224,6 @@ static int __init cpufreq_gov_userspace_init(void) ...@@ -206,7 +224,6 @@ static int __init cpufreq_gov_userspace_init(void)
static void __exit cpufreq_gov_userspace_exit(void) static void __exit cpufreq_gov_userspace_exit(void)
{ {
cpufreq_unregister_governor(&cpufreq_gov_userspace); cpufreq_unregister_governor(&cpufreq_gov_userspace);
cpufreq_unregister_notifier(&userspace_cpufreq_notifier_block, CPUFREQ_TRANSITION_NOTIFIER);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment