Commit 4ded3835 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'x86-fixes-for-linus' of...

Merge branch 'x86-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/x86/linux-2.6-tip

* 'x86-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/x86/linux-2.6-tip:
  x86, fpu: fix CONFIG_PREEMPT=y corruption of application's FPU stack
  suspend-vs-iommu: prevent suspend if we could not resume
  x86: section mismatch fix
  x86: fix Xorg crash with xf86MapVidMem error
  x86: fix pointer type warning in arch/x86/mm/init_64.c:early_memtest
  x86: fix bad pmd ffff810000207xxx(9090909090909090)
  x86: ioremap fix failing nesting check
  x86: fix broken math-emu with lazy allocation of fpu area
  x86: enable preemption in delay
  x86: disable preemption in native_smp_prepare_cpus
  x86: fix APIC warning on 32bit v2
parents e97dcb0e 870568b3
...@@ -242,12 +242,19 @@ static int __init acpi_parse_madt(struct acpi_table_header *table) ...@@ -242,12 +242,19 @@ static int __init acpi_parse_madt(struct acpi_table_header *table)
static void __cpuinit acpi_register_lapic(int id, u8 enabled) static void __cpuinit acpi_register_lapic(int id, u8 enabled)
{ {
unsigned int ver = 0;
if (!enabled) { if (!enabled) {
++disabled_cpus; ++disabled_cpus;
return; return;
} }
generic_processor_info(id, 0); #ifdef CONFIG_X86_32
if (boot_cpu_physical_apicid != -1U)
ver = apic_version[boot_cpu_physical_apicid];
#endif
generic_processor_info(id, ver);
} }
static int __init static int __init
...@@ -767,8 +774,13 @@ static void __init acpi_register_lapic_address(unsigned long address) ...@@ -767,8 +774,13 @@ static void __init acpi_register_lapic_address(unsigned long address)
mp_lapic_addr = address; mp_lapic_addr = address;
set_fixmap_nocache(FIX_APIC_BASE, address); set_fixmap_nocache(FIX_APIC_BASE, address);
if (boot_cpu_physical_apicid == -1U) if (boot_cpu_physical_apicid == -1U) {
boot_cpu_physical_apicid = GET_APIC_ID(read_apic_id()); boot_cpu_physical_apicid = GET_APIC_ID(read_apic_id());
#ifdef CONFIG_X86_32
apic_version[boot_cpu_physical_apicid] =
GET_APIC_VERSION(apic_read(APIC_LVR));
#endif
}
} }
static int __init early_acpi_parse_madt_lapic_addr_ovr(void) static int __init early_acpi_parse_madt_lapic_addr_ovr(void)
......
...@@ -56,6 +56,11 @@ void __cpuinit mxcsr_feature_mask_init(void) ...@@ -56,6 +56,11 @@ void __cpuinit mxcsr_feature_mask_init(void)
void __init init_thread_xstate(void) void __init init_thread_xstate(void)
{ {
if (!HAVE_HWFP) {
xstate_size = sizeof(struct i387_soft_struct);
return;
}
if (cpu_has_fxsr) if (cpu_has_fxsr)
xstate_size = sizeof(struct i387_fxsave_struct); xstate_size = sizeof(struct i387_fxsave_struct);
#ifdef CONFIG_X86_32 #ifdef CONFIG_X86_32
...@@ -94,7 +99,7 @@ void __cpuinit fpu_init(void) ...@@ -94,7 +99,7 @@ void __cpuinit fpu_init(void)
int init_fpu(struct task_struct *tsk) int init_fpu(struct task_struct *tsk)
{ {
if (tsk_used_math(tsk)) { if (tsk_used_math(tsk)) {
if (tsk == current) if (HAVE_HWFP && tsk == current)
unlazy_fpu(tsk); unlazy_fpu(tsk);
return 0; return 0;
} }
...@@ -109,6 +114,15 @@ int init_fpu(struct task_struct *tsk) ...@@ -109,6 +114,15 @@ int init_fpu(struct task_struct *tsk)
return -ENOMEM; return -ENOMEM;
} }
#ifdef CONFIG_X86_32
if (!HAVE_HWFP) {
memset(tsk->thread.xstate, 0, xstate_size);
finit();
set_stopped_child_used_math(tsk);
return 0;
}
#endif
if (cpu_has_fxsr) { if (cpu_has_fxsr) {
struct i387_fxsave_struct *fx = &tsk->thread.xstate->fxsave; struct i387_fxsave_struct *fx = &tsk->thread.xstate->fxsave;
...@@ -330,13 +344,13 @@ int fpregs_get(struct task_struct *target, const struct user_regset *regset, ...@@ -330,13 +344,13 @@ int fpregs_get(struct task_struct *target, const struct user_regset *regset,
struct user_i387_ia32_struct env; struct user_i387_ia32_struct env;
int ret; int ret;
if (!HAVE_HWFP)
return fpregs_soft_get(target, regset, pos, count, kbuf, ubuf);
ret = init_fpu(target); ret = init_fpu(target);
if (ret) if (ret)
return ret; return ret;
if (!HAVE_HWFP)
return fpregs_soft_get(target, regset, pos, count, kbuf, ubuf);
if (!cpu_has_fxsr) { if (!cpu_has_fxsr) {
return user_regset_copyout(&pos, &count, &kbuf, &ubuf, return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
&target->thread.xstate->fsave, 0, &target->thread.xstate->fsave, 0,
...@@ -360,15 +374,15 @@ int fpregs_set(struct task_struct *target, const struct user_regset *regset, ...@@ -360,15 +374,15 @@ int fpregs_set(struct task_struct *target, const struct user_regset *regset,
struct user_i387_ia32_struct env; struct user_i387_ia32_struct env;
int ret; int ret;
if (!HAVE_HWFP)
return fpregs_soft_set(target, regset, pos, count, kbuf, ubuf);
ret = init_fpu(target); ret = init_fpu(target);
if (ret) if (ret)
return ret; return ret;
set_stopped_child_used_math(target); set_stopped_child_used_math(target);
if (!HAVE_HWFP)
return fpregs_soft_set(target, regset, pos, count, kbuf, ubuf);
if (!cpu_has_fxsr) { if (!cpu_has_fxsr) {
return user_regset_copyin(&pos, &count, &kbuf, &ubuf, return user_regset_copyin(&pos, &count, &kbuf, &ubuf,
&target->thread.xstate->fsave, 0, -1); &target->thread.xstate->fsave, 0, -1);
...@@ -474,18 +488,18 @@ static int restore_i387_fxsave(struct _fpstate_ia32 __user *buf) ...@@ -474,18 +488,18 @@ static int restore_i387_fxsave(struct _fpstate_ia32 __user *buf)
int restore_i387_ia32(struct _fpstate_ia32 __user *buf) int restore_i387_ia32(struct _fpstate_ia32 __user *buf)
{ {
int err; int err;
struct task_struct *tsk = current;
if (HAVE_HWFP) { if (HAVE_HWFP)
struct task_struct *tsk = current;
clear_fpu(tsk); clear_fpu(tsk);
if (!used_math()) { if (!used_math()) {
err = init_fpu(tsk); err = init_fpu(tsk);
if (err) if (err)
return err; return err;
} }
if (HAVE_HWFP) {
if (cpu_has_fxsr) if (cpu_has_fxsr)
err = restore_i387_fxsave(buf); err = restore_i387_fxsave(buf);
else else
......
...@@ -26,6 +26,7 @@ ...@@ -26,6 +26,7 @@
#include <linux/kdebug.h> #include <linux/kdebug.h>
#include <linux/scatterlist.h> #include <linux/scatterlist.h>
#include <linux/iommu-helper.h> #include <linux/iommu-helper.h>
#include <linux/sysdev.h>
#include <asm/atomic.h> #include <asm/atomic.h>
#include <asm/io.h> #include <asm/io.h>
#include <asm/mtrr.h> #include <asm/mtrr.h>
...@@ -548,6 +549,28 @@ static __init unsigned read_aperture(struct pci_dev *dev, u32 *size) ...@@ -548,6 +549,28 @@ static __init unsigned read_aperture(struct pci_dev *dev, u32 *size)
return aper_base; return aper_base;
} }
static int gart_resume(struct sys_device *dev)
{
return 0;
}
static int gart_suspend(struct sys_device *dev, pm_message_t state)
{
return -EINVAL;
}
static struct sysdev_class gart_sysdev_class = {
.name = "gart",
.suspend = gart_suspend,
.resume = gart_resume,
};
static struct sys_device device_gart = {
.id = 0,
.cls = &gart_sysdev_class,
};
/* /*
* Private Northbridge GATT initialization in case we cannot use the * Private Northbridge GATT initialization in case we cannot use the
* AGP driver for some reason. * AGP driver for some reason.
...@@ -558,7 +581,7 @@ static __init int init_k8_gatt(struct agp_kern_info *info) ...@@ -558,7 +581,7 @@ static __init int init_k8_gatt(struct agp_kern_info *info)
unsigned aper_base, new_aper_base; unsigned aper_base, new_aper_base;
struct pci_dev *dev; struct pci_dev *dev;
void *gatt; void *gatt;
int i; int i, error;
printk(KERN_INFO "PCI-DMA: Disabling AGP.\n"); printk(KERN_INFO "PCI-DMA: Disabling AGP.\n");
aper_size = aper_base = info->aper_size = 0; aper_size = aper_base = info->aper_size = 0;
...@@ -606,6 +629,12 @@ static __init int init_k8_gatt(struct agp_kern_info *info) ...@@ -606,6 +629,12 @@ static __init int init_k8_gatt(struct agp_kern_info *info)
pci_write_config_dword(dev, 0x90, ctl); pci_write_config_dword(dev, 0x90, ctl);
} }
error = sysdev_class_register(&gart_sysdev_class);
if (!error)
error = sysdev_register(&device_gart);
if (error)
panic("Could not register gart_sysdev -- would corrupt data on next suspend");
flush_gart(); flush_gart();
printk(KERN_INFO "PCI-DMA: aperture base @ %x size %u KB\n", printk(KERN_INFO "PCI-DMA: aperture base @ %x size %u KB\n",
......
...@@ -649,8 +649,11 @@ struct task_struct * __switch_to(struct task_struct *prev_p, struct task_struct ...@@ -649,8 +649,11 @@ struct task_struct * __switch_to(struct task_struct *prev_p, struct task_struct
/* If the task has used fpu the last 5 timeslices, just do a full /* If the task has used fpu the last 5 timeslices, just do a full
* restore of the math state immediately to avoid the trap; the * restore of the math state immediately to avoid the trap; the
* chances of needing FPU soon are obviously high now * chances of needing FPU soon are obviously high now
*
* tsk_used_math() checks prevent calling math_state_restore(),
* which can sleep in the case of !tsk_used_math()
*/ */
if (next_p->fpu_counter > 5) if (tsk_used_math(next_p) && next_p->fpu_counter > 5)
math_state_restore(); math_state_restore();
/* /*
......
...@@ -658,8 +658,11 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) ...@@ -658,8 +658,11 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p)
/* If the task has used fpu the last 5 timeslices, just do a full /* If the task has used fpu the last 5 timeslices, just do a full
* restore of the math state immediately to avoid the trap; the * restore of the math state immediately to avoid the trap; the
* chances of needing FPU soon are obviously high now * chances of needing FPU soon are obviously high now
*
* tsk_used_math() checks prevent calling math_state_restore(),
* which can sleep in the case of !tsk_used_math()
*/ */
if (next_p->fpu_counter>5) if (tsk_used_math(next_p) && next_p->fpu_counter > 5)
math_state_restore(); math_state_restore();
return prev_p; return prev_p;
} }
......
...@@ -1190,6 +1190,7 @@ static void __init smp_cpu_index_default(void) ...@@ -1190,6 +1190,7 @@ static void __init smp_cpu_index_default(void)
*/ */
void __init native_smp_prepare_cpus(unsigned int max_cpus) void __init native_smp_prepare_cpus(unsigned int max_cpus)
{ {
preempt_disable();
nmi_watchdog_default(); nmi_watchdog_default();
smp_cpu_index_default(); smp_cpu_index_default();
current_cpu_data = boot_cpu_data; current_cpu_data = boot_cpu_data;
...@@ -1206,7 +1207,7 @@ void __init native_smp_prepare_cpus(unsigned int max_cpus) ...@@ -1206,7 +1207,7 @@ void __init native_smp_prepare_cpus(unsigned int max_cpus)
if (smp_sanity_check(max_cpus) < 0) { if (smp_sanity_check(max_cpus) < 0) {
printk(KERN_INFO "SMP disabled\n"); printk(KERN_INFO "SMP disabled\n");
disable_smp(); disable_smp();
return; goto out;
} }
preempt_disable(); preempt_disable();
...@@ -1246,6 +1247,8 @@ void __init native_smp_prepare_cpus(unsigned int max_cpus) ...@@ -1246,6 +1247,8 @@ void __init native_smp_prepare_cpus(unsigned int max_cpus)
printk(KERN_INFO "CPU%d: ", 0); printk(KERN_INFO "CPU%d: ", 0);
print_cpu_info(&cpu_data(0)); print_cpu_info(&cpu_data(0));
setup_boot_clock(); setup_boot_clock();
out:
preempt_enable();
} }
/* /*
* Early setup to make printk work. * Early setup to make printk work.
......
...@@ -44,13 +44,36 @@ static void delay_loop(unsigned long loops) ...@@ -44,13 +44,36 @@ static void delay_loop(unsigned long loops)
static void delay_tsc(unsigned long loops) static void delay_tsc(unsigned long loops)
{ {
unsigned long bclock, now; unsigned long bclock, now;
int cpu;
preempt_disable(); /* TSC's are per-cpu */ preempt_disable();
cpu = smp_processor_id();
rdtscl(bclock); rdtscl(bclock);
do { for (;;) {
rep_nop();
rdtscl(now); rdtscl(now);
} while ((now-bclock) < loops); if ((now - bclock) >= loops)
break;
/* Allow RT tasks to run */
preempt_enable();
rep_nop();
preempt_disable();
/*
* It is possible that we moved to another CPU, and
* since TSC's are per-cpu we need to calculate
* that. The delay must guarantee that we wait "at
* least" the amount of time. Being moved to another
* CPU could make the wait longer but we just need to
* make sure we waited long enough. Rebalance the
* counter for this CPU.
*/
if (unlikely(cpu != smp_processor_id())) {
loops -= (now - bclock);
cpu = smp_processor_id();
rdtscl(bclock);
}
}
preempt_enable(); preempt_enable();
} }
......
...@@ -31,14 +31,36 @@ int __devinit read_current_timer(unsigned long *timer_value) ...@@ -31,14 +31,36 @@ int __devinit read_current_timer(unsigned long *timer_value)
void __delay(unsigned long loops) void __delay(unsigned long loops)
{ {
unsigned bclock, now; unsigned bclock, now;
int cpu;
preempt_disable(); /* TSC's are pre-cpu */ preempt_disable();
cpu = smp_processor_id();
rdtscl(bclock); rdtscl(bclock);
do { for (;;) {
rep_nop();
rdtscl(now); rdtscl(now);
if ((now - bclock) >= loops)
break;
/* Allow RT tasks to run */
preempt_enable();
rep_nop();
preempt_disable();
/*
* It is possible that we moved to another CPU, and
* since TSC's are per-cpu we need to calculate
* that. The delay must guarantee that we wait "at
* least" the amount of time. Being moved to another
* CPU could make the wait longer but we just need to
* make sure we waited long enough. Rebalance the
* counter for this CPU.
*/
if (unlikely(cpu != smp_processor_id())) {
loops -= (now - bclock);
cpu = smp_processor_id();
rdtscl(bclock);
}
} }
while ((now-bclock) < loops);
preempt_enable(); preempt_enable();
} }
EXPORT_SYMBOL(__delay); EXPORT_SYMBOL(__delay);
......
...@@ -30,6 +30,7 @@ ...@@ -30,6 +30,7 @@
#include <asm/uaccess.h> #include <asm/uaccess.h>
#include <asm/desc.h> #include <asm/desc.h>
#include <asm/user.h> #include <asm/user.h>
#include <asm/i387.h>
#include "fpu_system.h" #include "fpu_system.h"
#include "fpu_emu.h" #include "fpu_emu.h"
...@@ -146,6 +147,13 @@ asmlinkage void math_emulate(long arg) ...@@ -146,6 +147,13 @@ asmlinkage void math_emulate(long arg)
unsigned long code_limit = 0; /* Initialized to stop compiler warnings */ unsigned long code_limit = 0; /* Initialized to stop compiler warnings */
struct desc_struct code_descriptor; struct desc_struct code_descriptor;
if (!used_math()) {
if (init_fpu(current)) {
do_group_exit(SIGKILL);
return;
}
}
#ifdef RE_ENTRANT_CHECKING #ifdef RE_ENTRANT_CHECKING
if (emulating) { if (emulating) {
printk("ERROR: wm-FPU-emu is not RE-ENTRANT!\n"); printk("ERROR: wm-FPU-emu is not RE-ENTRANT!\n");
...@@ -153,11 +161,6 @@ asmlinkage void math_emulate(long arg) ...@@ -153,11 +161,6 @@ asmlinkage void math_emulate(long arg)
RE_ENTRANT_CHECK_ON; RE_ENTRANT_CHECK_ON;
#endif /* RE_ENTRANT_CHECKING */ #endif /* RE_ENTRANT_CHECKING */
if (!used_math()) {
finit();
set_used_math();
}
SETUP_DATA_AREA(arg); SETUP_DATA_AREA(arg);
FPU_ORIG_EIP = FPU_EIP; FPU_ORIG_EIP = FPU_EIP;
......
...@@ -206,7 +206,7 @@ void __init cleanup_highmap(void) ...@@ -206,7 +206,7 @@ void __init cleanup_highmap(void)
pmd_t *last_pmd = pmd + PTRS_PER_PMD; pmd_t *last_pmd = pmd + PTRS_PER_PMD;
for (; pmd < last_pmd; pmd++, vaddr += PMD_SIZE) { for (; pmd < last_pmd; pmd++, vaddr += PMD_SIZE) {
if (!pmd_present(*pmd)) if (pmd_none(*pmd))
continue; continue;
if (vaddr < (unsigned long) _text || vaddr > end) if (vaddr < (unsigned long) _text || vaddr > end)
set_pmd(pmd, __pmd(0)); set_pmd(pmd, __pmd(0));
...@@ -506,7 +506,7 @@ early_param("memtest", parse_memtest); ...@@ -506,7 +506,7 @@ early_param("memtest", parse_memtest);
static void __init early_memtest(unsigned long start, unsigned long end) static void __init early_memtest(unsigned long start, unsigned long end)
{ {
unsigned long t_start, t_size; u64 t_start, t_size;
unsigned pattern; unsigned pattern;
if (!memtest_pattern) if (!memtest_pattern)
...@@ -525,7 +525,7 @@ static void __init early_memtest(unsigned long start, unsigned long end) ...@@ -525,7 +525,7 @@ static void __init early_memtest(unsigned long start, unsigned long end)
if (t_start + t_size > end) if (t_start + t_size > end)
t_size = end - t_start; t_size = end - t_start;
printk(KERN_CONT "\n %016lx - %016lx pattern %d", printk(KERN_CONT "\n %016llx - %016llx pattern %d",
t_start, t_start + t_size, pattern); t_start, t_start + t_size, pattern);
memtest(t_start, t_size, pattern); memtest(t_start, t_size, pattern);
......
...@@ -593,10 +593,11 @@ void __init early_iounmap(void *addr, unsigned long size) ...@@ -593,10 +593,11 @@ void __init early_iounmap(void *addr, unsigned long size)
unsigned long offset; unsigned long offset;
unsigned int nrpages; unsigned int nrpages;
enum fixed_addresses idx; enum fixed_addresses idx;
unsigned int nesting; int nesting;
nesting = --early_ioremap_nested; nesting = --early_ioremap_nested;
WARN_ON(nesting < 0); if (WARN_ON(nesting < 0))
return;
if (early_ioremap_debug) { if (early_ioremap_debug) {
printk(KERN_INFO "early_iounmap(%p, %08lx) [%d]\n", addr, printk(KERN_INFO "early_iounmap(%p, %08lx) [%d]\n", addr,
......
...@@ -34,7 +34,7 @@ void __cpuinit pat_disable(char *reason) ...@@ -34,7 +34,7 @@ void __cpuinit pat_disable(char *reason)
printk(KERN_INFO "%s\n", reason); printk(KERN_INFO "%s\n", reason);
} }
static int nopat(char *str) static int __init nopat(char *str)
{ {
pat_disable("PAT support disabled."); pat_disable("PAT support disabled.");
return 0; return 0;
...@@ -151,32 +151,33 @@ static int pat_x_mtrr_type(u64 start, u64 end, unsigned long prot, ...@@ -151,32 +151,33 @@ static int pat_x_mtrr_type(u64 start, u64 end, unsigned long prot,
unsigned long pat_type; unsigned long pat_type;
u8 mtrr_type; u8 mtrr_type;
mtrr_type = mtrr_type_lookup(start, end);
if (mtrr_type == 0xFF) { /* MTRR not enabled */
*ret_prot = prot;
return 0;
}
if (mtrr_type == 0xFE) { /* MTRR match error */
*ret_prot = _PAGE_CACHE_UC;
return -1;
}
if (mtrr_type != MTRR_TYPE_UNCACHABLE &&
mtrr_type != MTRR_TYPE_WRBACK &&
mtrr_type != MTRR_TYPE_WRCOMB) { /* MTRR type unhandled */
*ret_prot = _PAGE_CACHE_UC;
return -1;
}
pat_type = prot & _PAGE_CACHE_MASK; pat_type = prot & _PAGE_CACHE_MASK;
prot &= (~_PAGE_CACHE_MASK); prot &= (~_PAGE_CACHE_MASK);
/* Currently doing intersection by hand. Optimize it later. */ /*
* We return the PAT request directly for types where PAT takes
* precedence with respect to MTRR and for UC_MINUS.
* Consistency checks with other PAT requests is done later
* while going through memtype list.
*/
if (pat_type == _PAGE_CACHE_WC) { if (pat_type == _PAGE_CACHE_WC) {
*ret_prot = prot | _PAGE_CACHE_WC; *ret_prot = prot | _PAGE_CACHE_WC;
return 0;
} else if (pat_type == _PAGE_CACHE_UC_MINUS) { } else if (pat_type == _PAGE_CACHE_UC_MINUS) {
*ret_prot = prot | _PAGE_CACHE_UC_MINUS; *ret_prot = prot | _PAGE_CACHE_UC_MINUS;
} else if (pat_type == _PAGE_CACHE_UC || return 0;
mtrr_type == MTRR_TYPE_UNCACHABLE) { } else if (pat_type == _PAGE_CACHE_UC) {
*ret_prot = prot | _PAGE_CACHE_UC;
return 0;
}
/*
* Look for MTRR hint to get the effective type in case where PAT
* request is for WB.
*/
mtrr_type = mtrr_type_lookup(start, end);
if (mtrr_type == MTRR_TYPE_UNCACHABLE) {
*ret_prot = prot | _PAGE_CACHE_UC; *ret_prot = prot | _PAGE_CACHE_UC;
} else if (mtrr_type == MTRR_TYPE_WRCOMB) { } else if (mtrr_type == MTRR_TYPE_WRCOMB) {
*ret_prot = prot | _PAGE_CACHE_WC; *ret_prot = prot | _PAGE_CACHE_WC;
...@@ -233,14 +234,12 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type, ...@@ -233,14 +234,12 @@ int reserve_memtype(u64 start, u64 end, unsigned long req_type,
if (req_type == -1) { if (req_type == -1) {
/* /*
* Special case where caller wants to inherit from mtrr or * Call mtrr_lookup to get the type hint. This is an
* existing pat mapping, defaulting to UC_MINUS in case of * optimization for /dev/mem mmap'ers into WB memory (BIOS
* no match. * tools and ACPI tools). Use WB request for WB memory and use
* UC_MINUS otherwise.
*/ */
u8 mtrr_type = mtrr_type_lookup(start, end); u8 mtrr_type = mtrr_type_lookup(start, end);
if (mtrr_type == 0xFE) { /* MTRR match error */
err = -1;
}
if (mtrr_type == MTRR_TYPE_WRBACK) { if (mtrr_type == MTRR_TYPE_WRBACK) {
req_type = _PAGE_CACHE_WB; req_type = _PAGE_CACHE_WB;
......
...@@ -193,6 +193,8 @@ static inline int restore_i387(struct _fpstate __user *buf) ...@@ -193,6 +193,8 @@ static inline int restore_i387(struct _fpstate __user *buf)
#else /* CONFIG_X86_32 */ #else /* CONFIG_X86_32 */
extern void finit(void);
static inline void tolerant_fwait(void) static inline void tolerant_fwait(void)
{ {
asm volatile("fnclex ; fwait"); asm volatile("fnclex ; fwait");
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment