Commit 77d43164 authored by Linus Torvalds's avatar Linus Torvalds

Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/sparc

Pull sparc fixes from David Miller:
 "Sparc perf stack traversal fixes from David Ahern"

* git://git.kernel.org/pub/scm/linux/kernel/git/davem/sparc:
  sparc64: perf: Use UREG_FP rather than UREG_I6
  sparc64: perf: Add sanity checking on addresses in user stack
  sparc64: Convert BUG_ON to warning
  sparc: perf: Disable pagefaults while walking userspace stacks
parents 55a7d4b8 f01cae4e
...@@ -49,6 +49,28 @@ do { \ ...@@ -49,6 +49,28 @@ do { \
__asm__ __volatile__ ("wr %%g0, %0, %%asi" : : "r" ((val).seg)); \ __asm__ __volatile__ ("wr %%g0, %0, %%asi" : : "r" ((val).seg)); \
} while(0) } while(0)
/*
* Test whether a block of memory is a valid user space address.
* Returns 0 if the range is valid, nonzero otherwise.
*/
static inline bool __chk_range_not_ok(unsigned long addr, unsigned long size, unsigned long limit)
{
if (__builtin_constant_p(size))
return addr > limit - size;
addr += size;
if (addr < size)
return true;
return addr > limit;
}
#define __range_not_ok(addr, size, limit) \
({ \
__chk_user_ptr(addr); \
__chk_range_not_ok((unsigned long __force)(addr), size, limit); \
})
static inline int __access_ok(const void __user * addr, unsigned long size) static inline int __access_ok(const void __user * addr, unsigned long size)
{ {
return 1; return 1;
......
...@@ -21,7 +21,7 @@ ...@@ -21,7 +21,7 @@
#include <asm/stacktrace.h> #include <asm/stacktrace.h>
#include <asm/cpudata.h> #include <asm/cpudata.h>
#include <asm/uaccess.h> #include <linux/uaccess.h>
#include <linux/atomic.h> #include <linux/atomic.h>
#include <asm/nmi.h> #include <asm/nmi.h>
#include <asm/pcr.h> #include <asm/pcr.h>
...@@ -1741,18 +1741,31 @@ void perf_callchain_kernel(struct perf_callchain_entry *entry, ...@@ -1741,18 +1741,31 @@ void perf_callchain_kernel(struct perf_callchain_entry *entry,
} while (entry->nr < PERF_MAX_STACK_DEPTH); } while (entry->nr < PERF_MAX_STACK_DEPTH);
} }
static inline int
valid_user_frame(const void __user *fp, unsigned long size)
{
/* addresses should be at least 4-byte aligned */
if (((unsigned long) fp) & 3)
return 0;
return (__range_not_ok(fp, size, TASK_SIZE) == 0);
}
static void perf_callchain_user_64(struct perf_callchain_entry *entry, static void perf_callchain_user_64(struct perf_callchain_entry *entry,
struct pt_regs *regs) struct pt_regs *regs)
{ {
unsigned long ufp; unsigned long ufp;
ufp = regs->u_regs[UREG_I6] + STACK_BIAS; ufp = regs->u_regs[UREG_FP] + STACK_BIAS;
do { do {
struct sparc_stackf __user *usf; struct sparc_stackf __user *usf;
struct sparc_stackf sf; struct sparc_stackf sf;
unsigned long pc; unsigned long pc;
usf = (struct sparc_stackf __user *)ufp; usf = (struct sparc_stackf __user *)ufp;
if (!valid_user_frame(usf, sizeof(sf)))
break;
if (__copy_from_user_inatomic(&sf, usf, sizeof(sf))) if (__copy_from_user_inatomic(&sf, usf, sizeof(sf)))
break; break;
...@@ -1767,7 +1780,7 @@ static void perf_callchain_user_32(struct perf_callchain_entry *entry, ...@@ -1767,7 +1780,7 @@ static void perf_callchain_user_32(struct perf_callchain_entry *entry,
{ {
unsigned long ufp; unsigned long ufp;
ufp = regs->u_regs[UREG_I6] & 0xffffffffUL; ufp = regs->u_regs[UREG_FP] & 0xffffffffUL;
do { do {
unsigned long pc; unsigned long pc;
...@@ -1803,8 +1816,13 @@ perf_callchain_user(struct perf_callchain_entry *entry, struct pt_regs *regs) ...@@ -1803,8 +1816,13 @@ perf_callchain_user(struct perf_callchain_entry *entry, struct pt_regs *regs)
return; return;
flushw_user(); flushw_user();
pagefault_disable();
if (test_thread_flag(TIF_32BIT)) if (test_thread_flag(TIF_32BIT))
perf_callchain_user_32(entry, regs); perf_callchain_user_32(entry, regs);
else else
perf_callchain_user_64(entry, regs); perf_callchain_user_64(entry, regs);
pagefault_enable();
} }
...@@ -413,8 +413,9 @@ asmlinkage void __kprobes do_sparc64_fault(struct pt_regs *regs) ...@@ -413,8 +413,9 @@ asmlinkage void __kprobes do_sparc64_fault(struct pt_regs *regs)
* that here. * that here.
*/ */
if ((fault_code & FAULT_CODE_ITLB) && !(vma->vm_flags & VM_EXEC)) { if ((fault_code & FAULT_CODE_ITLB) && !(vma->vm_flags & VM_EXEC)) {
BUG_ON(address != regs->tpc); WARN(address != regs->tpc,
BUG_ON(regs->tstate & TSTATE_PRIV); "address (%lx) != regs->tpc (%lx)\n", address, regs->tpc);
WARN_ON(regs->tstate & TSTATE_PRIV);
goto bad_area; goto bad_area;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment