Commit 72825454 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'x86-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip

Pull x86 fixes from Ingo Molnar:
 "Misc fixes all over the place:

   - might_sleep() atomicity fix in the microcode loader

   - resctrl boundary condition fix

   - APIC arithmethics bug fix for frequencies >= 4.2 GHz

   - three 5-level paging crash fixes

   - two speculation fixes

   - a perf/stacktrace fix"

* 'x86-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip:
  x86/unwind/orc: Fall back to using frame pointers for generated code
  perf/x86: Always store regs->ip in perf_callchain_kernel()
  x86/speculation: Allow guests to use SSBD even if host does not
  x86/mm: Handle physical-virtual alignment mismatch in phys_p4d_init()
  x86/boot/64: Add missing fixup_pointer() for next_early_pgt access
  x86/boot/64: Fix crash if kernel image crosses page table boundary
  x86/apic: Fix integer overflow on 10 bit left shift of cpu_khz
  x86/resctrl: Prevent possible overrun during bitmap operations
  x86/microcode: Fix the microcode load on CPU hotplug for real
parents 57103eb7 ae6a45a0
...@@ -2402,13 +2402,13 @@ perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *re ...@@ -2402,13 +2402,13 @@ perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *re
return; return;
} }
if (perf_hw_regs(regs)) { if (perf_callchain_store(entry, regs->ip))
if (perf_callchain_store(entry, regs->ip)) return;
return;
if (perf_hw_regs(regs))
unwind_start(&state, current, regs, NULL); unwind_start(&state, current, regs, NULL);
} else { else
unwind_start(&state, current, NULL, (void *)regs->sp); unwind_start(&state, current, NULL, (void *)regs->sp);
}
for (; !unwind_done(&state); unwind_next_frame(&state)) { for (; !unwind_done(&state); unwind_next_frame(&state)) {
addr = unwind_get_return_address(&state); addr = unwind_get_return_address(&state);
......
...@@ -1464,7 +1464,8 @@ static void apic_pending_intr_clear(void) ...@@ -1464,7 +1464,8 @@ static void apic_pending_intr_clear(void)
if (queued) { if (queued) {
if (boot_cpu_has(X86_FEATURE_TSC) && cpu_khz) { if (boot_cpu_has(X86_FEATURE_TSC) && cpu_khz) {
ntsc = rdtsc(); ntsc = rdtsc();
max_loops = (cpu_khz << 10) - (ntsc - tsc); max_loops = (long long)cpu_khz << 10;
max_loops -= ntsc - tsc;
} else { } else {
max_loops--; max_loops--;
} }
......
...@@ -835,6 +835,16 @@ static enum ssb_mitigation __init __ssb_select_mitigation(void) ...@@ -835,6 +835,16 @@ static enum ssb_mitigation __init __ssb_select_mitigation(void)
break; break;
} }
/*
* If SSBD is controlled by the SPEC_CTRL MSR, then set the proper
* bit in the mask to allow guests to use the mitigation even in the
* case where the host does not enable it.
*/
if (static_cpu_has(X86_FEATURE_SPEC_CTRL_SSBD) ||
static_cpu_has(X86_FEATURE_AMD_SSBD)) {
x86_spec_ctrl_mask |= SPEC_CTRL_SSBD;
}
/* /*
* We have three CPU feature flags that are in play here: * We have three CPU feature flags that are in play here:
* - X86_BUG_SPEC_STORE_BYPASS - CPU is susceptible. * - X86_BUG_SPEC_STORE_BYPASS - CPU is susceptible.
...@@ -852,7 +862,6 @@ static enum ssb_mitigation __init __ssb_select_mitigation(void) ...@@ -852,7 +862,6 @@ static enum ssb_mitigation __init __ssb_select_mitigation(void)
x86_amd_ssb_disable(); x86_amd_ssb_disable();
} else { } else {
x86_spec_ctrl_base |= SPEC_CTRL_SSBD; x86_spec_ctrl_base |= SPEC_CTRL_SSBD;
x86_spec_ctrl_mask |= SPEC_CTRL_SSBD;
wrmsrl(MSR_IA32_SPEC_CTRL, x86_spec_ctrl_base); wrmsrl(MSR_IA32_SPEC_CTRL, x86_spec_ctrl_base);
} }
} }
......
...@@ -789,13 +789,16 @@ static struct syscore_ops mc_syscore_ops = { ...@@ -789,13 +789,16 @@ static struct syscore_ops mc_syscore_ops = {
.resume = mc_bp_resume, .resume = mc_bp_resume,
}; };
static int mc_cpu_online(unsigned int cpu) static int mc_cpu_starting(unsigned int cpu)
{ {
struct device *dev;
dev = get_cpu_device(cpu);
microcode_update_cpu(cpu); microcode_update_cpu(cpu);
pr_debug("CPU%d added\n", cpu); pr_debug("CPU%d added\n", cpu);
return 0;
}
static int mc_cpu_online(unsigned int cpu)
{
struct device *dev = get_cpu_device(cpu);
if (sysfs_create_group(&dev->kobj, &mc_attr_group)) if (sysfs_create_group(&dev->kobj, &mc_attr_group))
pr_err("Failed to create group for CPU%d\n", cpu); pr_err("Failed to create group for CPU%d\n", cpu);
...@@ -872,7 +875,9 @@ int __init microcode_init(void) ...@@ -872,7 +875,9 @@ int __init microcode_init(void)
goto out_ucode_group; goto out_ucode_group;
register_syscore_ops(&mc_syscore_ops); register_syscore_ops(&mc_syscore_ops);
cpuhp_setup_state_nocalls(CPUHP_AP_MICROCODE_LOADER, "x86/microcode:online", cpuhp_setup_state_nocalls(CPUHP_AP_MICROCODE_LOADER, "x86/microcode:starting",
mc_cpu_starting, NULL);
cpuhp_setup_state_nocalls(CPUHP_AP_ONLINE_DYN, "x86/microcode:online",
mc_cpu_online, mc_cpu_down_prep); mc_cpu_online, mc_cpu_down_prep);
pr_info("Microcode Update Driver: v%s.", DRIVER_VERSION); pr_info("Microcode Update Driver: v%s.", DRIVER_VERSION);
......
...@@ -796,8 +796,12 @@ static int rdt_bit_usage_show(struct kernfs_open_file *of, ...@@ -796,8 +796,12 @@ static int rdt_bit_usage_show(struct kernfs_open_file *of,
struct seq_file *seq, void *v) struct seq_file *seq, void *v)
{ {
struct rdt_resource *r = of->kn->parent->priv; struct rdt_resource *r = of->kn->parent->priv;
u32 sw_shareable = 0, hw_shareable = 0; /*
u32 exclusive = 0, pseudo_locked = 0; * Use unsigned long even though only 32 bits are used to ensure
* test_bit() is used safely.
*/
unsigned long sw_shareable = 0, hw_shareable = 0;
unsigned long exclusive = 0, pseudo_locked = 0;
struct rdt_domain *dom; struct rdt_domain *dom;
int i, hwb, swb, excl, psl; int i, hwb, swb, excl, psl;
enum rdtgrp_mode mode; enum rdtgrp_mode mode;
...@@ -842,10 +846,10 @@ static int rdt_bit_usage_show(struct kernfs_open_file *of, ...@@ -842,10 +846,10 @@ static int rdt_bit_usage_show(struct kernfs_open_file *of,
} }
for (i = r->cache.cbm_len - 1; i >= 0; i--) { for (i = r->cache.cbm_len - 1; i >= 0; i--) {
pseudo_locked = dom->plr ? dom->plr->cbm : 0; pseudo_locked = dom->plr ? dom->plr->cbm : 0;
hwb = test_bit(i, (unsigned long *)&hw_shareable); hwb = test_bit(i, &hw_shareable);
swb = test_bit(i, (unsigned long *)&sw_shareable); swb = test_bit(i, &sw_shareable);
excl = test_bit(i, (unsigned long *)&exclusive); excl = test_bit(i, &exclusive);
psl = test_bit(i, (unsigned long *)&pseudo_locked); psl = test_bit(i, &pseudo_locked);
if (hwb && swb) if (hwb && swb)
seq_putc(seq, 'X'); seq_putc(seq, 'X');
else if (hwb && !swb) else if (hwb && !swb)
...@@ -2486,26 +2490,19 @@ static int mkdir_mondata_all(struct kernfs_node *parent_kn, ...@@ -2486,26 +2490,19 @@ static int mkdir_mondata_all(struct kernfs_node *parent_kn,
*/ */
static void cbm_ensure_valid(u32 *_val, struct rdt_resource *r) static void cbm_ensure_valid(u32 *_val, struct rdt_resource *r)
{ {
/* unsigned long val = *_val;
* Convert the u32 _val to an unsigned long required by all the bit
* operations within this function. No more than 32 bits of this
* converted value can be accessed because all bit operations are
* additionally provided with cbm_len that is initialized during
* hardware enumeration using five bits from the EAX register and
* thus never can exceed 32 bits.
*/
unsigned long *val = (unsigned long *)_val;
unsigned int cbm_len = r->cache.cbm_len; unsigned int cbm_len = r->cache.cbm_len;
unsigned long first_bit, zero_bit; unsigned long first_bit, zero_bit;
if (*val == 0) if (val == 0)
return; return;
first_bit = find_first_bit(val, cbm_len); first_bit = find_first_bit(&val, cbm_len);
zero_bit = find_next_zero_bit(val, cbm_len, first_bit); zero_bit = find_next_zero_bit(&val, cbm_len, first_bit);
/* Clear any remaining bits to ensure contiguous region */ /* Clear any remaining bits to ensure contiguous region */
bitmap_clear(val, zero_bit, cbm_len - zero_bit); bitmap_clear(&val, zero_bit, cbm_len - zero_bit);
*_val = (u32)val;
} }
/* /*
......
...@@ -184,24 +184,25 @@ unsigned long __head __startup_64(unsigned long physaddr, ...@@ -184,24 +184,25 @@ unsigned long __head __startup_64(unsigned long physaddr,
pgtable_flags = _KERNPG_TABLE_NOENC + sme_get_me_mask(); pgtable_flags = _KERNPG_TABLE_NOENC + sme_get_me_mask();
if (la57) { if (la57) {
p4d = fixup_pointer(early_dynamic_pgts[next_early_pgt++], physaddr); p4d = fixup_pointer(early_dynamic_pgts[(*next_pgt_ptr)++],
physaddr);
i = (physaddr >> PGDIR_SHIFT) % PTRS_PER_PGD; i = (physaddr >> PGDIR_SHIFT) % PTRS_PER_PGD;
pgd[i + 0] = (pgdval_t)p4d + pgtable_flags; pgd[i + 0] = (pgdval_t)p4d + pgtable_flags;
pgd[i + 1] = (pgdval_t)p4d + pgtable_flags; pgd[i + 1] = (pgdval_t)p4d + pgtable_flags;
i = (physaddr >> P4D_SHIFT) % PTRS_PER_P4D; i = physaddr >> P4D_SHIFT;
p4d[i + 0] = (pgdval_t)pud + pgtable_flags; p4d[(i + 0) % PTRS_PER_P4D] = (pgdval_t)pud + pgtable_flags;
p4d[i + 1] = (pgdval_t)pud + pgtable_flags; p4d[(i + 1) % PTRS_PER_P4D] = (pgdval_t)pud + pgtable_flags;
} else { } else {
i = (physaddr >> PGDIR_SHIFT) % PTRS_PER_PGD; i = (physaddr >> PGDIR_SHIFT) % PTRS_PER_PGD;
pgd[i + 0] = (pgdval_t)pud + pgtable_flags; pgd[i + 0] = (pgdval_t)pud + pgtable_flags;
pgd[i + 1] = (pgdval_t)pud + pgtable_flags; pgd[i + 1] = (pgdval_t)pud + pgtable_flags;
} }
i = (physaddr >> PUD_SHIFT) % PTRS_PER_PUD; i = physaddr >> PUD_SHIFT;
pud[i + 0] = (pudval_t)pmd + pgtable_flags; pud[(i + 0) % PTRS_PER_PUD] = (pudval_t)pmd + pgtable_flags;
pud[i + 1] = (pudval_t)pmd + pgtable_flags; pud[(i + 1) % PTRS_PER_PUD] = (pudval_t)pmd + pgtable_flags;
pmd_entry = __PAGE_KERNEL_LARGE_EXEC & ~_PAGE_GLOBAL; pmd_entry = __PAGE_KERNEL_LARGE_EXEC & ~_PAGE_GLOBAL;
/* Filter out unsupported __PAGE_KERNEL_* bits: */ /* Filter out unsupported __PAGE_KERNEL_* bits: */
...@@ -211,8 +212,9 @@ unsigned long __head __startup_64(unsigned long physaddr, ...@@ -211,8 +212,9 @@ unsigned long __head __startup_64(unsigned long physaddr,
pmd_entry += physaddr; pmd_entry += physaddr;
for (i = 0; i < DIV_ROUND_UP(_end - _text, PMD_SIZE); i++) { for (i = 0; i < DIV_ROUND_UP(_end - _text, PMD_SIZE); i++) {
int idx = i + (physaddr >> PMD_SHIFT) % PTRS_PER_PMD; int idx = i + (physaddr >> PMD_SHIFT);
pmd[idx] = pmd_entry + i * PMD_SIZE;
pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE;
} }
/* /*
......
...@@ -82,9 +82,9 @@ static struct orc_entry *orc_find(unsigned long ip); ...@@ -82,9 +82,9 @@ static struct orc_entry *orc_find(unsigned long ip);
* But they are copies of the ftrace entries that are static and * But they are copies of the ftrace entries that are static and
* defined in ftrace_*.S, which do have orc entries. * defined in ftrace_*.S, which do have orc entries.
* *
* If the undwinder comes across a ftrace trampoline, then find the * If the unwinder comes across a ftrace trampoline, then find the
* ftrace function that was used to create it, and use that ftrace * ftrace function that was used to create it, and use that ftrace
* function's orc entrie, as the placement of the return code in * function's orc entry, as the placement of the return code in
* the stack will be identical. * the stack will be identical.
*/ */
static struct orc_entry *orc_ftrace_find(unsigned long ip) static struct orc_entry *orc_ftrace_find(unsigned long ip)
...@@ -128,6 +128,16 @@ static struct orc_entry null_orc_entry = { ...@@ -128,6 +128,16 @@ static struct orc_entry null_orc_entry = {
.type = ORC_TYPE_CALL .type = ORC_TYPE_CALL
}; };
/* Fake frame pointer entry -- used as a fallback for generated code */
static struct orc_entry orc_fp_entry = {
.type = ORC_TYPE_CALL,
.sp_reg = ORC_REG_BP,
.sp_offset = 16,
.bp_reg = ORC_REG_PREV_SP,
.bp_offset = -16,
.end = 0,
};
static struct orc_entry *orc_find(unsigned long ip) static struct orc_entry *orc_find(unsigned long ip)
{ {
static struct orc_entry *orc; static struct orc_entry *orc;
...@@ -392,8 +402,16 @@ bool unwind_next_frame(struct unwind_state *state) ...@@ -392,8 +402,16 @@ bool unwind_next_frame(struct unwind_state *state)
* calls and calls to noreturn functions. * calls and calls to noreturn functions.
*/ */
orc = orc_find(state->signal ? state->ip : state->ip - 1); orc = orc_find(state->signal ? state->ip : state->ip - 1);
if (!orc) if (!orc) {
goto err; /*
* As a fallback, try to assume this code uses a frame pointer.
* This is useful for generated code, like BPF, which ORC
* doesn't know about. This is just a guess, so the rest of
* the unwind is no longer considered reliable.
*/
orc = &orc_fp_entry;
state->error = true;
}
/* End-of-stack check for kernel threads: */ /* End-of-stack check for kernel threads: */
if (orc->sp_reg == ORC_REG_UNDEFINED) { if (orc->sp_reg == ORC_REG_UNDEFINED) {
......
...@@ -671,23 +671,25 @@ static unsigned long __meminit ...@@ -671,23 +671,25 @@ static unsigned long __meminit
phys_p4d_init(p4d_t *p4d_page, unsigned long paddr, unsigned long paddr_end, phys_p4d_init(p4d_t *p4d_page, unsigned long paddr, unsigned long paddr_end,
unsigned long page_size_mask, bool init) unsigned long page_size_mask, bool init)
{ {
unsigned long paddr_next, paddr_last = paddr_end; unsigned long vaddr, vaddr_end, vaddr_next, paddr_next, paddr_last;
unsigned long vaddr = (unsigned long)__va(paddr);
int i = p4d_index(vaddr); paddr_last = paddr_end;
vaddr = (unsigned long)__va(paddr);
vaddr_end = (unsigned long)__va(paddr_end);
if (!pgtable_l5_enabled()) if (!pgtable_l5_enabled())
return phys_pud_init((pud_t *) p4d_page, paddr, paddr_end, return phys_pud_init((pud_t *) p4d_page, paddr, paddr_end,
page_size_mask, init); page_size_mask, init);
for (; i < PTRS_PER_P4D; i++, paddr = paddr_next) { for (; vaddr < vaddr_end; vaddr = vaddr_next) {
p4d_t *p4d; p4d_t *p4d = p4d_page + p4d_index(vaddr);
pud_t *pud; pud_t *pud;
vaddr = (unsigned long)__va(paddr); vaddr_next = (vaddr & P4D_MASK) + P4D_SIZE;
p4d = p4d_page + p4d_index(vaddr); paddr = __pa(vaddr);
paddr_next = (paddr & P4D_MASK) + P4D_SIZE;
if (paddr >= paddr_end) { if (paddr >= paddr_end) {
paddr_next = __pa(vaddr_next);
if (!after_bootmem && if (!after_bootmem &&
!e820__mapped_any(paddr & P4D_MASK, paddr_next, !e820__mapped_any(paddr & P4D_MASK, paddr_next,
E820_TYPE_RAM) && E820_TYPE_RAM) &&
...@@ -699,13 +701,13 @@ phys_p4d_init(p4d_t *p4d_page, unsigned long paddr, unsigned long paddr_end, ...@@ -699,13 +701,13 @@ phys_p4d_init(p4d_t *p4d_page, unsigned long paddr, unsigned long paddr_end,
if (!p4d_none(*p4d)) { if (!p4d_none(*p4d)) {
pud = pud_offset(p4d, 0); pud = pud_offset(p4d, 0);
paddr_last = phys_pud_init(pud, paddr, paddr_end, paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end),
page_size_mask, init); page_size_mask, init);
continue; continue;
} }
pud = alloc_low_page(); pud = alloc_low_page();
paddr_last = phys_pud_init(pud, paddr, paddr_end, paddr_last = phys_pud_init(pud, paddr, __pa(vaddr_end),
page_size_mask, init); page_size_mask, init);
spin_lock(&init_mm.page_table_lock); spin_lock(&init_mm.page_table_lock);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment