Commit 43c1266c authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'perfcounters-rename-for-linus' of...

Merge branch 'perfcounters-rename-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip

* 'perfcounters-rename-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
  perf: Tidy up after the big rename
  perf: Do the big rename: Performance Counters -> Performance Events
  perf_counter: Rename 'event' to event_id/hw_event
  perf_counter: Rename list_entry -> group_entry, counter_list -> group_list

Manually resolved some fairly trivial conflicts with the tracing tree in
include/trace/ftrace.h and kernel/trace/trace_syscalls.c.
parents b8c7f1dc 57c0c15b
...@@ -4000,7 +4000,7 @@ S: Maintained ...@@ -4000,7 +4000,7 @@ S: Maintained
F: include/linux/delayacct.h F: include/linux/delayacct.h
F: kernel/delayacct.c F: kernel/delayacct.c
PERFORMANCE COUNTER SUBSYSTEM PERFORMANCE EVENTS SUBSYSTEM
M: Peter Zijlstra <a.p.zijlstra@chello.nl> M: Peter Zijlstra <a.p.zijlstra@chello.nl>
M: Paul Mackerras <paulus@samba.org> M: Paul Mackerras <paulus@samba.org>
M: Ingo Molnar <mingo@elte.hu> M: Ingo Molnar <mingo@elte.hu>
......
...@@ -390,7 +390,7 @@ ...@@ -390,7 +390,7 @@
#define __NR_preadv (__NR_SYSCALL_BASE+361) #define __NR_preadv (__NR_SYSCALL_BASE+361)
#define __NR_pwritev (__NR_SYSCALL_BASE+362) #define __NR_pwritev (__NR_SYSCALL_BASE+362)
#define __NR_rt_tgsigqueueinfo (__NR_SYSCALL_BASE+363) #define __NR_rt_tgsigqueueinfo (__NR_SYSCALL_BASE+363)
#define __NR_perf_counter_open (__NR_SYSCALL_BASE+364) #define __NR_perf_event_open (__NR_SYSCALL_BASE+364)
/* /*
* The following SWIs are ARM private. * The following SWIs are ARM private.
......
...@@ -373,7 +373,7 @@ ...@@ -373,7 +373,7 @@
CALL(sys_preadv) CALL(sys_preadv)
CALL(sys_pwritev) CALL(sys_pwritev)
CALL(sys_rt_tgsigqueueinfo) CALL(sys_rt_tgsigqueueinfo)
CALL(sys_perf_counter_open) CALL(sys_perf_event_open)
#ifndef syscalls_counted #ifndef syscalls_counted
.equ syscalls_padding, ((NR_syscalls + 3) & ~3) - NR_syscalls .equ syscalls_padding, ((NR_syscalls + 3) & ~3) - NR_syscalls
#define syscalls_counted #define syscalls_counted
......
...@@ -381,7 +381,7 @@ ...@@ -381,7 +381,7 @@
#define __NR_preadv 366 #define __NR_preadv 366
#define __NR_pwritev 367 #define __NR_pwritev 367
#define __NR_rt_tgsigqueueinfo 368 #define __NR_rt_tgsigqueueinfo 368
#define __NR_perf_counter_open 369 #define __NR_perf_event_open 369
#define __NR_syscall 370 #define __NR_syscall 370
#define NR_syscalls __NR_syscall #define NR_syscalls __NR_syscall
......
...@@ -1620,7 +1620,7 @@ ENTRY(_sys_call_table) ...@@ -1620,7 +1620,7 @@ ENTRY(_sys_call_table)
.long _sys_preadv .long _sys_preadv
.long _sys_pwritev .long _sys_pwritev
.long _sys_rt_tgsigqueueinfo .long _sys_rt_tgsigqueueinfo
.long _sys_perf_counter_open .long _sys_perf_event_open
.rept NR_syscalls-(.-_sys_call_table)/4 .rept NR_syscalls-(.-_sys_call_table)/4
.long _sys_ni_syscall .long _sys_ni_syscall
......
...@@ -7,7 +7,7 @@ config FRV ...@@ -7,7 +7,7 @@ config FRV
default y default y
select HAVE_IDE select HAVE_IDE
select HAVE_ARCH_TRACEHOOK select HAVE_ARCH_TRACEHOOK
select HAVE_PERF_COUNTERS select HAVE_PERF_EVENTS
config ZONE_DMA config ZONE_DMA
bool bool
......
/* FRV performance counter support /* FRV performance event support
* *
* Copyright (C) 2009 Red Hat, Inc. All Rights Reserved. * Copyright (C) 2009 Red Hat, Inc. All Rights Reserved.
* Written by David Howells (dhowells@redhat.com) * Written by David Howells (dhowells@redhat.com)
...@@ -9,9 +9,9 @@ ...@@ -9,9 +9,9 @@
* 2 of the Licence, or (at your option) any later version. * 2 of the Licence, or (at your option) any later version.
*/ */
#ifndef _ASM_PERF_COUNTER_H #ifndef _ASM_PERF_EVENT_H
#define _ASM_PERF_COUNTER_H #define _ASM_PERF_EVENT_H
#define PERF_COUNTER_INDEX_OFFSET 0 #define PERF_EVENT_INDEX_OFFSET 0
#endif /* _ASM_PERF_COUNTER_H */ #endif /* _ASM_PERF_EVENT_H */
...@@ -342,7 +342,7 @@ ...@@ -342,7 +342,7 @@
#define __NR_preadv 333 #define __NR_preadv 333
#define __NR_pwritev 334 #define __NR_pwritev 334
#define __NR_rt_tgsigqueueinfo 335 #define __NR_rt_tgsigqueueinfo 335
#define __NR_perf_counter_open 336 #define __NR_perf_event_open 336
#ifdef __KERNEL__ #ifdef __KERNEL__
......
...@@ -1525,6 +1525,6 @@ sys_call_table: ...@@ -1525,6 +1525,6 @@ sys_call_table:
.long sys_preadv .long sys_preadv
.long sys_pwritev .long sys_pwritev
.long sys_rt_tgsigqueueinfo /* 335 */ .long sys_rt_tgsigqueueinfo /* 335 */
.long sys_perf_counter_open .long sys_perf_event_open
syscall_table_size = (. - sys_call_table) syscall_table_size = (. - sys_call_table)
...@@ -5,4 +5,4 @@ ...@@ -5,4 +5,4 @@
lib-y := \ lib-y := \
__ashldi3.o __lshrdi3.o __muldi3.o __ashrdi3.o __negdi2.o __ucmpdi2.o \ __ashldi3.o __lshrdi3.o __muldi3.o __ashrdi3.o __negdi2.o __ucmpdi2.o \
checksum.o memcpy.o memset.o atomic-ops.o atomic64-ops.o \ checksum.o memcpy.o memset.o atomic-ops.o atomic64-ops.o \
outsl_ns.o outsl_sw.o insl_ns.o insl_sw.o cache.o perf_counter.o outsl_ns.o outsl_sw.o insl_ns.o insl_sw.o cache.o perf_event.o
/* Performance counter handling /* Performance event handling
* *
* Copyright (C) 2009 Red Hat, Inc. All Rights Reserved. * Copyright (C) 2009 Red Hat, Inc. All Rights Reserved.
* Written by David Howells (dhowells@redhat.com) * Written by David Howells (dhowells@redhat.com)
...@@ -9,11 +9,11 @@ ...@@ -9,11 +9,11 @@
* 2 of the Licence, or (at your option) any later version. * 2 of the Licence, or (at your option) any later version.
*/ */
#include <linux/perf_counter.h> #include <linux/perf_event.h>
/* /*
* mark the performance counter as pending * mark the performance event as pending
*/ */
void set_perf_counter_pending(void) void set_perf_event_pending(void)
{ {
} }
...@@ -335,7 +335,7 @@ ...@@ -335,7 +335,7 @@
#define __NR_preadv 329 #define __NR_preadv 329
#define __NR_pwritev 330 #define __NR_pwritev 330
#define __NR_rt_tgsigqueueinfo 331 #define __NR_rt_tgsigqueueinfo 331
#define __NR_perf_counter_open 332 #define __NR_perf_event_open 332
#ifdef __KERNEL__ #ifdef __KERNEL__
......
...@@ -756,5 +756,5 @@ sys_call_table: ...@@ -756,5 +756,5 @@ sys_call_table:
.long sys_preadv .long sys_preadv
.long sys_pwritev /* 330 */ .long sys_pwritev /* 330 */
.long sys_rt_tgsigqueueinfo .long sys_rt_tgsigqueueinfo
.long sys_perf_counter_open .long sys_perf_event_open
...@@ -350,7 +350,7 @@ ENTRY(sys_call_table) ...@@ -350,7 +350,7 @@ ENTRY(sys_call_table)
.long sys_preadv .long sys_preadv
.long sys_pwritev /* 330 */ .long sys_pwritev /* 330 */
.long sys_rt_tgsigqueueinfo .long sys_rt_tgsigqueueinfo
.long sys_perf_counter_open .long sys_perf_event_open
.rept NR_syscalls-(.-sys_call_table)/4 .rept NR_syscalls-(.-sys_call_table)/4
.long sys_ni_syscall .long sys_ni_syscall
......
...@@ -381,7 +381,7 @@ ...@@ -381,7 +381,7 @@
#define __NR_preadv 363 /* new */ #define __NR_preadv 363 /* new */
#define __NR_pwritev 364 /* new */ #define __NR_pwritev 364 /* new */
#define __NR_rt_tgsigqueueinfo 365 /* new */ #define __NR_rt_tgsigqueueinfo 365 /* new */
#define __NR_perf_counter_open 366 /* new */ #define __NR_perf_event_open 366 /* new */
#define __NR_syscalls 367 #define __NR_syscalls 367
......
...@@ -370,4 +370,4 @@ ENTRY(sys_call_table) ...@@ -370,4 +370,4 @@ ENTRY(sys_call_table)
.long sys_ni_syscall .long sys_ni_syscall
.long sys_ni_syscall .long sys_ni_syscall
.long sys_rt_tgsigqueueinfo /* 365 */ .long sys_rt_tgsigqueueinfo /* 365 */
.long sys_perf_counter_open .long sys_perf_event_open
...@@ -353,7 +353,7 @@ ...@@ -353,7 +353,7 @@
#define __NR_preadv (__NR_Linux + 330) #define __NR_preadv (__NR_Linux + 330)
#define __NR_pwritev (__NR_Linux + 331) #define __NR_pwritev (__NR_Linux + 331)
#define __NR_rt_tgsigqueueinfo (__NR_Linux + 332) #define __NR_rt_tgsigqueueinfo (__NR_Linux + 332)
#define __NR_perf_counter_open (__NR_Linux + 333) #define __NR_perf_event_open (__NR_Linux + 333)
#define __NR_accept4 (__NR_Linux + 334) #define __NR_accept4 (__NR_Linux + 334)
/* /*
...@@ -664,7 +664,7 @@ ...@@ -664,7 +664,7 @@
#define __NR_preadv (__NR_Linux + 289) #define __NR_preadv (__NR_Linux + 289)
#define __NR_pwritev (__NR_Linux + 290) #define __NR_pwritev (__NR_Linux + 290)
#define __NR_rt_tgsigqueueinfo (__NR_Linux + 291) #define __NR_rt_tgsigqueueinfo (__NR_Linux + 291)
#define __NR_perf_counter_open (__NR_Linux + 292) #define __NR_perf_event_open (__NR_Linux + 292)
#define __NR_accept4 (__NR_Linux + 293) #define __NR_accept4 (__NR_Linux + 293)
/* /*
...@@ -979,7 +979,7 @@ ...@@ -979,7 +979,7 @@
#define __NR_preadv (__NR_Linux + 293) #define __NR_preadv (__NR_Linux + 293)
#define __NR_pwritev (__NR_Linux + 294) #define __NR_pwritev (__NR_Linux + 294)
#define __NR_rt_tgsigqueueinfo (__NR_Linux + 295) #define __NR_rt_tgsigqueueinfo (__NR_Linux + 295)
#define __NR_perf_counter_open (__NR_Linux + 296) #define __NR_perf_event_open (__NR_Linux + 296)
#define __NR_accept4 (__NR_Linux + 297) #define __NR_accept4 (__NR_Linux + 297)
/* /*
......
...@@ -581,7 +581,7 @@ einval: li v0, -ENOSYS ...@@ -581,7 +581,7 @@ einval: li v0, -ENOSYS
sys sys_preadv 6 /* 4330 */ sys sys_preadv 6 /* 4330 */
sys sys_pwritev 6 sys sys_pwritev 6
sys sys_rt_tgsigqueueinfo 4 sys sys_rt_tgsigqueueinfo 4
sys sys_perf_counter_open 5 sys sys_perf_event_open 5
sys sys_accept4 4 sys sys_accept4 4
.endm .endm
......
...@@ -418,6 +418,6 @@ sys_call_table: ...@@ -418,6 +418,6 @@ sys_call_table:
PTR sys_preadv PTR sys_preadv
PTR sys_pwritev /* 5390 */ PTR sys_pwritev /* 5390 */
PTR sys_rt_tgsigqueueinfo PTR sys_rt_tgsigqueueinfo
PTR sys_perf_counter_open PTR sys_perf_event_open
PTR sys_accept4 PTR sys_accept4
.size sys_call_table,.-sys_call_table .size sys_call_table,.-sys_call_table
...@@ -416,6 +416,6 @@ EXPORT(sysn32_call_table) ...@@ -416,6 +416,6 @@ EXPORT(sysn32_call_table)
PTR sys_preadv PTR sys_preadv
PTR sys_pwritev PTR sys_pwritev
PTR compat_sys_rt_tgsigqueueinfo /* 5295 */ PTR compat_sys_rt_tgsigqueueinfo /* 5295 */
PTR sys_perf_counter_open PTR sys_perf_event_open
PTR sys_accept4 PTR sys_accept4
.size sysn32_call_table,.-sysn32_call_table .size sysn32_call_table,.-sysn32_call_table
...@@ -536,6 +536,6 @@ sys_call_table: ...@@ -536,6 +536,6 @@ sys_call_table:
PTR compat_sys_preadv /* 4330 */ PTR compat_sys_preadv /* 4330 */
PTR compat_sys_pwritev PTR compat_sys_pwritev
PTR compat_sys_rt_tgsigqueueinfo PTR compat_sys_rt_tgsigqueueinfo
PTR sys_perf_counter_open PTR sys_perf_event_open
PTR sys_accept4 PTR sys_accept4
.size sys_call_table,.-sys_call_table .size sys_call_table,.-sys_call_table
...@@ -347,7 +347,7 @@ ...@@ -347,7 +347,7 @@
#define __NR_preadv 334 #define __NR_preadv 334
#define __NR_pwritev 335 #define __NR_pwritev 335
#define __NR_rt_tgsigqueueinfo 336 #define __NR_rt_tgsigqueueinfo 336
#define __NR_perf_counter_open 337 #define __NR_perf_event_open 337
#ifdef __KERNEL__ #ifdef __KERNEL__
......
...@@ -723,7 +723,7 @@ ENTRY(sys_call_table) ...@@ -723,7 +723,7 @@ ENTRY(sys_call_table)
.long sys_preadv .long sys_preadv
.long sys_pwritev /* 335 */ .long sys_pwritev /* 335 */
.long sys_rt_tgsigqueueinfo .long sys_rt_tgsigqueueinfo
.long sys_perf_counter_open .long sys_perf_event_open
nr_syscalls=(.-sys_call_table)/4 nr_syscalls=(.-sys_call_table)/4
...@@ -16,7 +16,7 @@ config PARISC ...@@ -16,7 +16,7 @@ config PARISC
select RTC_DRV_GENERIC select RTC_DRV_GENERIC
select INIT_ALL_POSSIBLE select INIT_ALL_POSSIBLE
select BUG select BUG
select HAVE_PERF_COUNTERS select HAVE_PERF_EVENTS
select GENERIC_ATOMIC64 if !64BIT select GENERIC_ATOMIC64 if !64BIT
help help
The PA-RISC microprocessor is designed by Hewlett-Packard and used The PA-RISC microprocessor is designed by Hewlett-Packard and used
......
#ifndef __ASM_PARISC_PERF_COUNTER_H
#define __ASM_PARISC_PERF_COUNTER_H
/* parisc only supports software counters through this interface. */
static inline void set_perf_counter_pending(void) { }
#endif /* __ASM_PARISC_PERF_COUNTER_H */
#ifndef __ASM_PARISC_PERF_EVENT_H
#define __ASM_PARISC_PERF_EVENT_H
/* parisc only supports software events through this interface. */
static inline void set_perf_event_pending(void) { }
#endif /* __ASM_PARISC_PERF_EVENT_H */
...@@ -810,9 +810,9 @@ ...@@ -810,9 +810,9 @@
#define __NR_preadv (__NR_Linux + 315) #define __NR_preadv (__NR_Linux + 315)
#define __NR_pwritev (__NR_Linux + 316) #define __NR_pwritev (__NR_Linux + 316)
#define __NR_rt_tgsigqueueinfo (__NR_Linux + 317) #define __NR_rt_tgsigqueueinfo (__NR_Linux + 317)
#define __NR_perf_counter_open (__NR_Linux + 318) #define __NR_perf_event_open (__NR_Linux + 318)
#define __NR_Linux_syscalls (__NR_perf_counter_open + 1) #define __NR_Linux_syscalls (__NR_perf_event_open + 1)
#define __IGNORE_select /* newselect */ #define __IGNORE_select /* newselect */
......
...@@ -416,7 +416,7 @@ ...@@ -416,7 +416,7 @@
ENTRY_COMP(preadv) /* 315 */ ENTRY_COMP(preadv) /* 315 */
ENTRY_COMP(pwritev) ENTRY_COMP(pwritev)
ENTRY_COMP(rt_tgsigqueueinfo) ENTRY_COMP(rt_tgsigqueueinfo)
ENTRY_SAME(perf_counter_open) ENTRY_SAME(perf_event_open)
/* Nothing yet */ /* Nothing yet */
......
...@@ -129,7 +129,7 @@ config PPC ...@@ -129,7 +129,7 @@ config PPC
select HAVE_OPROFILE select HAVE_OPROFILE
select HAVE_SYSCALL_WRAPPERS if PPC64 select HAVE_SYSCALL_WRAPPERS if PPC64
select GENERIC_ATOMIC64 if PPC32 select GENERIC_ATOMIC64 if PPC32
select HAVE_PERF_COUNTERS select HAVE_PERF_EVENTS
config EARLY_PRINTK config EARLY_PRINTK
bool bool
......
...@@ -135,43 +135,43 @@ static inline int irqs_disabled_flags(unsigned long flags) ...@@ -135,43 +135,43 @@ static inline int irqs_disabled_flags(unsigned long flags)
*/ */
struct irq_chip; struct irq_chip;
#ifdef CONFIG_PERF_COUNTERS #ifdef CONFIG_PERF_EVENTS
#ifdef CONFIG_PPC64 #ifdef CONFIG_PPC64
static inline unsigned long test_perf_counter_pending(void) static inline unsigned long test_perf_event_pending(void)
{ {
unsigned long x; unsigned long x;
asm volatile("lbz %0,%1(13)" asm volatile("lbz %0,%1(13)"
: "=r" (x) : "=r" (x)
: "i" (offsetof(struct paca_struct, perf_counter_pending))); : "i" (offsetof(struct paca_struct, perf_event_pending)));
return x; return x;
} }
static inline void set_perf_counter_pending(void) static inline void set_perf_event_pending(void)
{ {
asm volatile("stb %0,%1(13)" : : asm volatile("stb %0,%1(13)" : :
"r" (1), "r" (1),
"i" (offsetof(struct paca_struct, perf_counter_pending))); "i" (offsetof(struct paca_struct, perf_event_pending)));
} }
static inline void clear_perf_counter_pending(void) static inline void clear_perf_event_pending(void)
{ {
asm volatile("stb %0,%1(13)" : : asm volatile("stb %0,%1(13)" : :
"r" (0), "r" (0),
"i" (offsetof(struct paca_struct, perf_counter_pending))); "i" (offsetof(struct paca_struct, perf_event_pending)));
} }
#endif /* CONFIG_PPC64 */ #endif /* CONFIG_PPC64 */
#else /* CONFIG_PERF_COUNTERS */ #else /* CONFIG_PERF_EVENTS */
static inline unsigned long test_perf_counter_pending(void) static inline unsigned long test_perf_event_pending(void)
{ {
return 0; return 0;
} }
static inline void clear_perf_counter_pending(void) {} static inline void clear_perf_event_pending(void) {}
#endif /* CONFIG_PERF_COUNTERS */ #endif /* CONFIG_PERF_EVENTS */
#endif /* __KERNEL__ */ #endif /* __KERNEL__ */
#endif /* _ASM_POWERPC_HW_IRQ_H */ #endif /* _ASM_POWERPC_HW_IRQ_H */
...@@ -122,7 +122,7 @@ struct paca_struct { ...@@ -122,7 +122,7 @@ struct paca_struct {
u8 soft_enabled; /* irq soft-enable flag */ u8 soft_enabled; /* irq soft-enable flag */
u8 hard_enabled; /* set if irqs are enabled in MSR */ u8 hard_enabled; /* set if irqs are enabled in MSR */
u8 io_sync; /* writel() needs spin_unlock sync */ u8 io_sync; /* writel() needs spin_unlock sync */
u8 perf_counter_pending; /* PM interrupt while soft-disabled */ u8 perf_event_pending; /* PM interrupt while soft-disabled */
/* Stuff for accurate time accounting */ /* Stuff for accurate time accounting */
u64 user_time; /* accumulated usermode TB ticks */ u64 user_time; /* accumulated usermode TB ticks */
......
/* /*
* Performance counter support - PowerPC-specific definitions. * Performance event support - PowerPC-specific definitions.
* *
* Copyright 2008-2009 Paul Mackerras, IBM Corporation. * Copyright 2008-2009 Paul Mackerras, IBM Corporation.
* *
...@@ -12,9 +12,9 @@ ...@@ -12,9 +12,9 @@
#include <asm/hw_irq.h> #include <asm/hw_irq.h>
#define MAX_HWCOUNTERS 8 #define MAX_HWEVENTS 8
#define MAX_EVENT_ALTERNATIVES 8 #define MAX_EVENT_ALTERNATIVES 8
#define MAX_LIMITED_HWCOUNTERS 2 #define MAX_LIMITED_HWEVENTS 2
/* /*
* This struct provides the constants and functions needed to * This struct provides the constants and functions needed to
...@@ -22,18 +22,18 @@ ...@@ -22,18 +22,18 @@
*/ */
struct power_pmu { struct power_pmu {
const char *name; const char *name;
int n_counter; int n_event;
int max_alternatives; int max_alternatives;
unsigned long add_fields; unsigned long add_fields;
unsigned long test_adder; unsigned long test_adder;
int (*compute_mmcr)(u64 events[], int n_ev, int (*compute_mmcr)(u64 events[], int n_ev,
unsigned int hwc[], unsigned long mmcr[]); unsigned int hwc[], unsigned long mmcr[]);
int (*get_constraint)(u64 event, unsigned long *mskp, int (*get_constraint)(u64 event_id, unsigned long *mskp,
unsigned long *valp); unsigned long *valp);
int (*get_alternatives)(u64 event, unsigned int flags, int (*get_alternatives)(u64 event_id, unsigned int flags,
u64 alt[]); u64 alt[]);
void (*disable_pmc)(unsigned int pmc, unsigned long mmcr[]); void (*disable_pmc)(unsigned int pmc, unsigned long mmcr[]);
int (*limited_pmc_event)(u64 event); int (*limited_pmc_event)(u64 event_id);
u32 flags; u32 flags;
int n_generic; int n_generic;
int *generic_events; int *generic_events;
...@@ -61,10 +61,10 @@ struct pt_regs; ...@@ -61,10 +61,10 @@ struct pt_regs;
extern unsigned long perf_misc_flags(struct pt_regs *regs); extern unsigned long perf_misc_flags(struct pt_regs *regs);
extern unsigned long perf_instruction_pointer(struct pt_regs *regs); extern unsigned long perf_instruction_pointer(struct pt_regs *regs);
#define PERF_COUNTER_INDEX_OFFSET 1 #define PERF_EVENT_INDEX_OFFSET 1
/* /*
* Only override the default definitions in include/linux/perf_counter.h * Only override the default definitions in include/linux/perf_event.h
* if we have hardware PMU support. * if we have hardware PMU support.
*/ */
#ifdef CONFIG_PPC_PERF_CTRS #ifdef CONFIG_PPC_PERF_CTRS
...@@ -73,14 +73,14 @@ extern unsigned long perf_instruction_pointer(struct pt_regs *regs); ...@@ -73,14 +73,14 @@ extern unsigned long perf_instruction_pointer(struct pt_regs *regs);
/* /*
* The power_pmu.get_constraint function returns a 32/64-bit value and * The power_pmu.get_constraint function returns a 32/64-bit value and
* a 32/64-bit mask that express the constraints between this event and * a 32/64-bit mask that express the constraints between this event_id and
* other events. * other events.
* *
* The value and mask are divided up into (non-overlapping) bitfields * The value and mask are divided up into (non-overlapping) bitfields
* of three different types: * of three different types:
* *
* Select field: this expresses the constraint that some set of bits * Select field: this expresses the constraint that some set of bits
* in MMCR* needs to be set to a specific value for this event. For a * in MMCR* needs to be set to a specific value for this event_id. For a
* select field, the mask contains 1s in every bit of the field, and * select field, the mask contains 1s in every bit of the field, and
* the value contains a unique value for each possible setting of the * the value contains a unique value for each possible setting of the
* MMCR* bits. The constraint checking code will ensure that two events * MMCR* bits. The constraint checking code will ensure that two events
...@@ -102,9 +102,9 @@ extern unsigned long perf_instruction_pointer(struct pt_regs *regs); ...@@ -102,9 +102,9 @@ extern unsigned long perf_instruction_pointer(struct pt_regs *regs);
* possible.) For N classes, the field is N+1 bits wide, and each class * possible.) For N classes, the field is N+1 bits wide, and each class
* is assigned one bit from the least-significant N bits. The mask has * is assigned one bit from the least-significant N bits. The mask has
* only the most-significant bit set, and the value has only the bit * only the most-significant bit set, and the value has only the bit
* for the event's class set. The test_adder has the least significant * for the event_id's class set. The test_adder has the least significant
* bit set in the field. * bit set in the field.
* *
* If an event is not subject to the constraint expressed by a particular * If an event_id is not subject to the constraint expressed by a particular
* field, then it will have 0 in both the mask and value for that field. * field, then it will have 0 in both the mask and value for that field.
*/ */
...@@ -322,7 +322,7 @@ SYSCALL_SPU(epoll_create1) ...@@ -322,7 +322,7 @@ SYSCALL_SPU(epoll_create1)
SYSCALL_SPU(dup3) SYSCALL_SPU(dup3)
SYSCALL_SPU(pipe2) SYSCALL_SPU(pipe2)
SYSCALL(inotify_init1) SYSCALL(inotify_init1)
SYSCALL_SPU(perf_counter_open) SYSCALL_SPU(perf_event_open)
COMPAT_SYS_SPU(preadv) COMPAT_SYS_SPU(preadv)
COMPAT_SYS_SPU(pwritev) COMPAT_SYS_SPU(pwritev)
COMPAT_SYS(rt_tgsigqueueinfo) COMPAT_SYS(rt_tgsigqueueinfo)
...@@ -341,7 +341,7 @@ ...@@ -341,7 +341,7 @@
#define __NR_dup3 316 #define __NR_dup3 316
#define __NR_pipe2 317 #define __NR_pipe2 317
#define __NR_inotify_init1 318 #define __NR_inotify_init1 318
#define __NR_perf_counter_open 319 #define __NR_perf_event_open 319
#define __NR_preadv 320 #define __NR_preadv 320
#define __NR_pwritev 321 #define __NR_pwritev 321
#define __NR_rt_tgsigqueueinfo 322 #define __NR_rt_tgsigqueueinfo 322
......
...@@ -97,7 +97,7 @@ obj64-$(CONFIG_AUDIT) += compat_audit.o ...@@ -97,7 +97,7 @@ obj64-$(CONFIG_AUDIT) += compat_audit.o
obj-$(CONFIG_DYNAMIC_FTRACE) += ftrace.o obj-$(CONFIG_DYNAMIC_FTRACE) += ftrace.o
obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += ftrace.o obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += ftrace.o
obj-$(CONFIG_PPC_PERF_CTRS) += perf_counter.o perf_callchain.o obj-$(CONFIG_PPC_PERF_CTRS) += perf_event.o perf_callchain.o
obj64-$(CONFIG_PPC_PERF_CTRS) += power4-pmu.o ppc970-pmu.o power5-pmu.o \ obj64-$(CONFIG_PPC_PERF_CTRS) += power4-pmu.o ppc970-pmu.o power5-pmu.o \
power5+-pmu.o power6-pmu.o power7-pmu.o power5+-pmu.o power6-pmu.o power7-pmu.o
obj32-$(CONFIG_PPC_PERF_CTRS) += mpc7450-pmu.o obj32-$(CONFIG_PPC_PERF_CTRS) += mpc7450-pmu.o
......
...@@ -133,7 +133,7 @@ int main(void) ...@@ -133,7 +133,7 @@ int main(void)
DEFINE(PACAKMSR, offsetof(struct paca_struct, kernel_msr)); DEFINE(PACAKMSR, offsetof(struct paca_struct, kernel_msr));
DEFINE(PACASOFTIRQEN, offsetof(struct paca_struct, soft_enabled)); DEFINE(PACASOFTIRQEN, offsetof(struct paca_struct, soft_enabled));
DEFINE(PACAHARDIRQEN, offsetof(struct paca_struct, hard_enabled)); DEFINE(PACAHARDIRQEN, offsetof(struct paca_struct, hard_enabled));
DEFINE(PACAPERFPEND, offsetof(struct paca_struct, perf_counter_pending)); DEFINE(PACAPERFPEND, offsetof(struct paca_struct, perf_event_pending));
DEFINE(PACACONTEXTID, offsetof(struct paca_struct, context.id)); DEFINE(PACACONTEXTID, offsetof(struct paca_struct, context.id));
#ifdef CONFIG_PPC_MM_SLICES #ifdef CONFIG_PPC_MM_SLICES
DEFINE(PACALOWSLICESPSIZE, offsetof(struct paca_struct, DEFINE(PACALOWSLICESPSIZE, offsetof(struct paca_struct,
......
...@@ -556,14 +556,14 @@ ALT_FW_FTR_SECTION_END_IFCLR(FW_FEATURE_ISERIES) ...@@ -556,14 +556,14 @@ ALT_FW_FTR_SECTION_END_IFCLR(FW_FEATURE_ISERIES)
2: 2:
TRACE_AND_RESTORE_IRQ(r5); TRACE_AND_RESTORE_IRQ(r5);
#ifdef CONFIG_PERF_COUNTERS #ifdef CONFIG_PERF_EVENTS
/* check paca->perf_counter_pending if we're enabling ints */ /* check paca->perf_event_pending if we're enabling ints */
lbz r3,PACAPERFPEND(r13) lbz r3,PACAPERFPEND(r13)
and. r3,r3,r5 and. r3,r3,r5
beq 27f beq 27f
bl .perf_counter_do_pending bl .perf_event_do_pending
27: 27:
#endif /* CONFIG_PERF_COUNTERS */ #endif /* CONFIG_PERF_EVENTS */
/* extract EE bit and use it to restore paca->hard_enabled */ /* extract EE bit and use it to restore paca->hard_enabled */
ld r3,_MSR(r1) ld r3,_MSR(r1)
......
...@@ -53,7 +53,7 @@ ...@@ -53,7 +53,7 @@
#include <linux/bootmem.h> #include <linux/bootmem.h>
#include <linux/pci.h> #include <linux/pci.h>
#include <linux/debugfs.h> #include <linux/debugfs.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <asm/uaccess.h> #include <asm/uaccess.h>
#include <asm/system.h> #include <asm/system.h>
...@@ -138,9 +138,9 @@ notrace void raw_local_irq_restore(unsigned long en) ...@@ -138,9 +138,9 @@ notrace void raw_local_irq_restore(unsigned long en)
} }
#endif /* CONFIG_PPC_STD_MMU_64 */ #endif /* CONFIG_PPC_STD_MMU_64 */
if (test_perf_counter_pending()) { if (test_perf_event_pending()) {
clear_perf_counter_pending(); clear_perf_event_pending();
perf_counter_do_pending(); perf_event_do_pending();
} }
/* /*
......
...@@ -9,7 +9,7 @@ ...@@ -9,7 +9,7 @@
* 2 of the License, or (at your option) any later version. * 2 of the License, or (at your option) any later version.
*/ */
#include <linux/string.h> #include <linux/string.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <asm/reg.h> #include <asm/reg.h>
#include <asm/cputable.h> #include <asm/cputable.h>
......
...@@ -10,7 +10,7 @@ ...@@ -10,7 +10,7 @@
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/sched.h> #include <linux/sched.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/percpu.h> #include <linux/percpu.h>
#include <linux/uaccess.h> #include <linux/uaccess.h>
#include <linux/mm.h> #include <linux/mm.h>
......
...@@ -9,7 +9,7 @@ ...@@ -9,7 +9,7 @@
* 2 of the License, or (at your option) any later version. * 2 of the License, or (at your option) any later version.
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/string.h> #include <linux/string.h>
#include <asm/reg.h> #include <asm/reg.h>
#include <asm/cputable.h> #include <asm/cputable.h>
......
...@@ -9,7 +9,7 @@ ...@@ -9,7 +9,7 @@
* 2 of the License, or (at your option) any later version. * 2 of the License, or (at your option) any later version.
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/string.h> #include <linux/string.h>
#include <asm/reg.h> #include <asm/reg.h>
#include <asm/cputable.h> #include <asm/cputable.h>
......
...@@ -9,7 +9,7 @@ ...@@ -9,7 +9,7 @@
* 2 of the License, or (at your option) any later version. * 2 of the License, or (at your option) any later version.
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/string.h> #include <linux/string.h>
#include <asm/reg.h> #include <asm/reg.h>
#include <asm/cputable.h> #include <asm/cputable.h>
......
...@@ -9,7 +9,7 @@ ...@@ -9,7 +9,7 @@
* 2 of the License, or (at your option) any later version. * 2 of the License, or (at your option) any later version.
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/string.h> #include <linux/string.h>
#include <asm/reg.h> #include <asm/reg.h>
#include <asm/cputable.h> #include <asm/cputable.h>
......
...@@ -9,7 +9,7 @@ ...@@ -9,7 +9,7 @@
* 2 of the License, or (at your option) any later version. * 2 of the License, or (at your option) any later version.
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/string.h> #include <linux/string.h>
#include <asm/reg.h> #include <asm/reg.h>
#include <asm/cputable.h> #include <asm/cputable.h>
......
...@@ -9,7 +9,7 @@ ...@@ -9,7 +9,7 @@
* 2 of the License, or (at your option) any later version. * 2 of the License, or (at your option) any later version.
*/ */
#include <linux/string.h> #include <linux/string.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <asm/reg.h> #include <asm/reg.h>
#include <asm/cputable.h> #include <asm/cputable.h>
......
...@@ -53,7 +53,7 @@ ...@@ -53,7 +53,7 @@
#include <linux/posix-timers.h> #include <linux/posix-timers.h>
#include <linux/irq.h> #include <linux/irq.h>
#include <linux/delay.h> #include <linux/delay.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <asm/io.h> #include <asm/io.h>
#include <asm/processor.h> #include <asm/processor.h>
...@@ -527,25 +527,25 @@ void __init iSeries_time_init_early(void) ...@@ -527,25 +527,25 @@ void __init iSeries_time_init_early(void)
} }
#endif /* CONFIG_PPC_ISERIES */ #endif /* CONFIG_PPC_ISERIES */
#if defined(CONFIG_PERF_COUNTERS) && defined(CONFIG_PPC32) #if defined(CONFIG_PERF_EVENTS) && defined(CONFIG_PPC32)
DEFINE_PER_CPU(u8, perf_counter_pending); DEFINE_PER_CPU(u8, perf_event_pending);
void set_perf_counter_pending(void) void set_perf_event_pending(void)
{ {
get_cpu_var(perf_counter_pending) = 1; get_cpu_var(perf_event_pending) = 1;
set_dec(1); set_dec(1);
put_cpu_var(perf_counter_pending); put_cpu_var(perf_event_pending);
} }
#define test_perf_counter_pending() __get_cpu_var(perf_counter_pending) #define test_perf_event_pending() __get_cpu_var(perf_event_pending)
#define clear_perf_counter_pending() __get_cpu_var(perf_counter_pending) = 0 #define clear_perf_event_pending() __get_cpu_var(perf_event_pending) = 0
#else /* CONFIG_PERF_COUNTERS && CONFIG_PPC32 */ #else /* CONFIG_PERF_EVENTS && CONFIG_PPC32 */
#define test_perf_counter_pending() 0 #define test_perf_event_pending() 0
#define clear_perf_counter_pending() #define clear_perf_event_pending()
#endif /* CONFIG_PERF_COUNTERS && CONFIG_PPC32 */ #endif /* CONFIG_PERF_EVENTS && CONFIG_PPC32 */
/* /*
* For iSeries shared processors, we have to let the hypervisor * For iSeries shared processors, we have to let the hypervisor
...@@ -573,9 +573,9 @@ void timer_interrupt(struct pt_regs * regs) ...@@ -573,9 +573,9 @@ void timer_interrupt(struct pt_regs * regs)
set_dec(DECREMENTER_MAX); set_dec(DECREMENTER_MAX);
#ifdef CONFIG_PPC32 #ifdef CONFIG_PPC32
if (test_perf_counter_pending()) { if (test_perf_event_pending()) {
clear_perf_counter_pending(); clear_perf_event_pending();
perf_counter_do_pending(); perf_event_do_pending();
} }
if (atomic_read(&ppc_n_lost_interrupts) != 0) if (atomic_read(&ppc_n_lost_interrupts) != 0)
do_IRQ(regs); do_IRQ(regs);
......
...@@ -29,7 +29,7 @@ ...@@ -29,7 +29,7 @@
#include <linux/module.h> #include <linux/module.h>
#include <linux/kprobes.h> #include <linux/kprobes.h>
#include <linux/kdebug.h> #include <linux/kdebug.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <asm/firmware.h> #include <asm/firmware.h>
#include <asm/page.h> #include <asm/page.h>
...@@ -171,7 +171,7 @@ int __kprobes do_page_fault(struct pt_regs *regs, unsigned long address, ...@@ -171,7 +171,7 @@ int __kprobes do_page_fault(struct pt_regs *regs, unsigned long address,
die("Weird page fault", regs, SIGSEGV); die("Weird page fault", regs, SIGSEGV);
} }
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address);
/* When running in the kernel we expect faults to occur only to /* When running in the kernel we expect faults to occur only to
* addresses in user space. All other faults represent errors in the * addresses in user space. All other faults represent errors in the
...@@ -312,7 +312,7 @@ int __kprobes do_page_fault(struct pt_regs *regs, unsigned long address, ...@@ -312,7 +312,7 @@ int __kprobes do_page_fault(struct pt_regs *regs, unsigned long address,
} }
if (ret & VM_FAULT_MAJOR) { if (ret & VM_FAULT_MAJOR) {
current->maj_flt++; current->maj_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0,
regs, address); regs, address);
#ifdef CONFIG_PPC_SMLPAR #ifdef CONFIG_PPC_SMLPAR
if (firmware_has_feature(FW_FEATURE_CMO)) { if (firmware_has_feature(FW_FEATURE_CMO)) {
...@@ -323,7 +323,7 @@ int __kprobes do_page_fault(struct pt_regs *regs, unsigned long address, ...@@ -323,7 +323,7 @@ int __kprobes do_page_fault(struct pt_regs *regs, unsigned long address,
#endif #endif
} else { } else {
current->min_flt++; current->min_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0,
regs, address); regs, address);
} }
up_read(&mm->mmap_sem); up_read(&mm->mmap_sem);
......
...@@ -280,9 +280,9 @@ config PPC_HAVE_PMU_SUPPORT ...@@ -280,9 +280,9 @@ config PPC_HAVE_PMU_SUPPORT
config PPC_PERF_CTRS config PPC_PERF_CTRS
def_bool y def_bool y
depends on PERF_COUNTERS && PPC_HAVE_PMU_SUPPORT depends on PERF_EVENTS && PPC_HAVE_PMU_SUPPORT
help help
This enables the powerpc-specific perf_counter back-end. This enables the powerpc-specific perf_event back-end.
config SMP config SMP
depends on PPC_BOOK3S || PPC_BOOK3E || FSL_BOOKE depends on PPC_BOOK3S || PPC_BOOK3E || FSL_BOOKE
......
...@@ -94,7 +94,7 @@ config S390 ...@@ -94,7 +94,7 @@ config S390
select HAVE_KVM if 64BIT select HAVE_KVM if 64BIT
select HAVE_ARCH_TRACEHOOK select HAVE_ARCH_TRACEHOOK
select INIT_ALL_POSSIBLE select INIT_ALL_POSSIBLE
select HAVE_PERF_COUNTERS select HAVE_PERF_EVENTS
config SCHED_OMIT_FRAME_POINTER config SCHED_OMIT_FRAME_POINTER
bool bool
......
/*
* Performance counter support - s390 specific definitions.
*
* Copyright 2009 Martin Schwidefsky, IBM Corporation.
*/
static inline void set_perf_counter_pending(void) {}
static inline void clear_perf_counter_pending(void) {}
#define PERF_COUNTER_INDEX_OFFSET 0
/*
* Performance event support - s390 specific definitions.
*
* Copyright 2009 Martin Schwidefsky, IBM Corporation.
*/
static inline void set_perf_event_pending(void) {}
static inline void clear_perf_event_pending(void) {}
#define PERF_EVENT_INDEX_OFFSET 0
...@@ -268,7 +268,7 @@ ...@@ -268,7 +268,7 @@
#define __NR_preadv 328 #define __NR_preadv 328
#define __NR_pwritev 329 #define __NR_pwritev 329
#define __NR_rt_tgsigqueueinfo 330 #define __NR_rt_tgsigqueueinfo 330
#define __NR_perf_counter_open 331 #define __NR_perf_event_open 331
#define NR_syscalls 332 #define NR_syscalls 332
/* /*
......
...@@ -1832,11 +1832,11 @@ compat_sys_rt_tgsigqueueinfo_wrapper: ...@@ -1832,11 +1832,11 @@ compat_sys_rt_tgsigqueueinfo_wrapper:
llgtr %r5,%r5 # struct compat_siginfo * llgtr %r5,%r5 # struct compat_siginfo *
jg compat_sys_rt_tgsigqueueinfo_wrapper # branch to system call jg compat_sys_rt_tgsigqueueinfo_wrapper # branch to system call
.globl sys_perf_counter_open_wrapper .globl sys_perf_event_open_wrapper
sys_perf_counter_open_wrapper: sys_perf_event_open_wrapper:
llgtr %r2,%r2 # const struct perf_counter_attr * llgtr %r2,%r2 # const struct perf_event_attr *
lgfr %r3,%r3 # pid_t lgfr %r3,%r3 # pid_t
lgfr %r4,%r4 # int lgfr %r4,%r4 # int
lgfr %r5,%r5 # int lgfr %r5,%r5 # int
llgfr %r6,%r6 # unsigned long llgfr %r6,%r6 # unsigned long
jg sys_perf_counter_open # branch to system call jg sys_perf_event_open # branch to system call
...@@ -339,4 +339,4 @@ SYSCALL(sys_epoll_create1,sys_epoll_create1,sys_epoll_create1_wrapper) ...@@ -339,4 +339,4 @@ SYSCALL(sys_epoll_create1,sys_epoll_create1,sys_epoll_create1_wrapper)
SYSCALL(sys_preadv,sys_preadv,compat_sys_preadv_wrapper) SYSCALL(sys_preadv,sys_preadv,compat_sys_preadv_wrapper)
SYSCALL(sys_pwritev,sys_pwritev,compat_sys_pwritev_wrapper) SYSCALL(sys_pwritev,sys_pwritev,compat_sys_pwritev_wrapper)
SYSCALL(sys_rt_tgsigqueueinfo,sys_rt_tgsigqueueinfo,compat_sys_rt_tgsigqueueinfo_wrapper) /* 330 */ SYSCALL(sys_rt_tgsigqueueinfo,sys_rt_tgsigqueueinfo,compat_sys_rt_tgsigqueueinfo_wrapper) /* 330 */
SYSCALL(sys_perf_counter_open,sys_perf_counter_open,sys_perf_counter_open_wrapper) SYSCALL(sys_perf_event_open,sys_perf_event_open,sys_perf_event_open_wrapper)
...@@ -10,7 +10,7 @@ ...@@ -10,7 +10,7 @@
* Copyright (C) 1995 Linus Torvalds * Copyright (C) 1995 Linus Torvalds
*/ */
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/signal.h> #include <linux/signal.h>
#include <linux/sched.h> #include <linux/sched.h>
#include <linux/kernel.h> #include <linux/kernel.h>
...@@ -306,7 +306,7 @@ do_exception(struct pt_regs *regs, unsigned long error_code, int write) ...@@ -306,7 +306,7 @@ do_exception(struct pt_regs *regs, unsigned long error_code, int write)
* interrupts again and then search the VMAs * interrupts again and then search the VMAs
*/ */
local_irq_enable(); local_irq_enable();
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address);
down_read(&mm->mmap_sem); down_read(&mm->mmap_sem);
si_code = SEGV_MAPERR; si_code = SEGV_MAPERR;
...@@ -366,11 +366,11 @@ do_exception(struct pt_regs *regs, unsigned long error_code, int write) ...@@ -366,11 +366,11 @@ do_exception(struct pt_regs *regs, unsigned long error_code, int write)
} }
if (fault & VM_FAULT_MAJOR) { if (fault & VM_FAULT_MAJOR) {
tsk->maj_flt++; tsk->maj_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0,
regs, address); regs, address);
} else { } else {
tsk->min_flt++; tsk->min_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0,
regs, address); regs, address);
} }
up_read(&mm->mmap_sem); up_read(&mm->mmap_sem);
......
...@@ -16,7 +16,7 @@ config SUPERH ...@@ -16,7 +16,7 @@ config SUPERH
select HAVE_IOREMAP_PROT if MMU select HAVE_IOREMAP_PROT if MMU
select HAVE_ARCH_TRACEHOOK select HAVE_ARCH_TRACEHOOK
select HAVE_DMA_API_DEBUG select HAVE_DMA_API_DEBUG
select HAVE_PERF_COUNTERS select HAVE_PERF_EVENTS
select HAVE_KERNEL_GZIP select HAVE_KERNEL_GZIP
select HAVE_KERNEL_BZIP2 select HAVE_KERNEL_BZIP2
select HAVE_KERNEL_LZMA select HAVE_KERNEL_LZMA
......
#ifndef __ASM_SH_PERF_COUNTER_H
#define __ASM_SH_PERF_COUNTER_H
/* SH only supports software counters through this interface. */
static inline void set_perf_counter_pending(void) {}
#define PERF_COUNTER_INDEX_OFFSET 0
#endif /* __ASM_SH_PERF_COUNTER_H */
#ifndef __ASM_SH_PERF_EVENT_H
#define __ASM_SH_PERF_EVENT_H
/* SH only supports software events through this interface. */
static inline void set_perf_event_pending(void) {}
#define PERF_EVENT_INDEX_OFFSET 0
#endif /* __ASM_SH_PERF_EVENT_H */
...@@ -344,7 +344,7 @@ ...@@ -344,7 +344,7 @@
#define __NR_preadv 333 #define __NR_preadv 333
#define __NR_pwritev 334 #define __NR_pwritev 334
#define __NR_rt_tgsigqueueinfo 335 #define __NR_rt_tgsigqueueinfo 335
#define __NR_perf_counter_open 336 #define __NR_perf_event_open 336
#define NR_syscalls 337 #define NR_syscalls 337
......
...@@ -384,7 +384,7 @@ ...@@ -384,7 +384,7 @@
#define __NR_preadv 361 #define __NR_preadv 361
#define __NR_pwritev 362 #define __NR_pwritev 362
#define __NR_rt_tgsigqueueinfo 363 #define __NR_rt_tgsigqueueinfo 363
#define __NR_perf_counter_open 364 #define __NR_perf_event_open 364
#ifdef __KERNEL__ #ifdef __KERNEL__
......
...@@ -352,4 +352,4 @@ ENTRY(sys_call_table) ...@@ -352,4 +352,4 @@ ENTRY(sys_call_table)
.long sys_preadv .long sys_preadv
.long sys_pwritev .long sys_pwritev
.long sys_rt_tgsigqueueinfo /* 335 */ .long sys_rt_tgsigqueueinfo /* 335 */
.long sys_perf_counter_open .long sys_perf_event_open
...@@ -390,4 +390,4 @@ sys_call_table: ...@@ -390,4 +390,4 @@ sys_call_table:
.long sys_preadv .long sys_preadv
.long sys_pwritev .long sys_pwritev
.long sys_rt_tgsigqueueinfo .long sys_rt_tgsigqueueinfo
.long sys_perf_counter_open .long sys_perf_event_open
...@@ -15,7 +15,7 @@ ...@@ -15,7 +15,7 @@
#include <linux/mm.h> #include <linux/mm.h>
#include <linux/hardirq.h> #include <linux/hardirq.h>
#include <linux/kprobes.h> #include <linux/kprobes.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <asm/io_trapped.h> #include <asm/io_trapped.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/mmu_context.h> #include <asm/mmu_context.h>
...@@ -157,7 +157,7 @@ asmlinkage void __kprobes do_page_fault(struct pt_regs *regs, ...@@ -157,7 +157,7 @@ asmlinkage void __kprobes do_page_fault(struct pt_regs *regs,
if ((regs->sr & SR_IMASK) != SR_IMASK) if ((regs->sr & SR_IMASK) != SR_IMASK)
local_irq_enable(); local_irq_enable();
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address);
/* /*
* If we're in an interrupt, have no user context or are running * If we're in an interrupt, have no user context or are running
...@@ -208,11 +208,11 @@ asmlinkage void __kprobes do_page_fault(struct pt_regs *regs, ...@@ -208,11 +208,11 @@ asmlinkage void __kprobes do_page_fault(struct pt_regs *regs,
} }
if (fault & VM_FAULT_MAJOR) { if (fault & VM_FAULT_MAJOR) {
tsk->maj_flt++; tsk->maj_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0,
regs, address); regs, address);
} else { } else {
tsk->min_flt++; tsk->min_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0,
regs, address); regs, address);
} }
......
...@@ -20,7 +20,7 @@ ...@@ -20,7 +20,7 @@
#include <linux/mman.h> #include <linux/mman.h>
#include <linux/mm.h> #include <linux/mm.h>
#include <linux/smp.h> #include <linux/smp.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/interrupt.h> #include <linux/interrupt.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/io.h> #include <asm/io.h>
...@@ -116,7 +116,7 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long writeaccess, ...@@ -116,7 +116,7 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long writeaccess,
/* Not an IO address, so reenable interrupts */ /* Not an IO address, so reenable interrupts */
local_irq_enable(); local_irq_enable();
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address);
/* /*
* If we're in an interrupt or have no user * If we're in an interrupt or have no user
...@@ -201,11 +201,11 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long writeaccess, ...@@ -201,11 +201,11 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long writeaccess,
if (fault & VM_FAULT_MAJOR) { if (fault & VM_FAULT_MAJOR) {
tsk->maj_flt++; tsk->maj_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0,
regs, address); regs, address);
} else { } else {
tsk->min_flt++; tsk->min_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0,
regs, address); regs, address);
} }
......
...@@ -25,7 +25,7 @@ config SPARC ...@@ -25,7 +25,7 @@ config SPARC
select ARCH_WANT_OPTIONAL_GPIOLIB select ARCH_WANT_OPTIONAL_GPIOLIB
select RTC_CLASS select RTC_CLASS
select RTC_DRV_M48T59 select RTC_DRV_M48T59
select HAVE_PERF_COUNTERS select HAVE_PERF_EVENTS
select HAVE_DMA_ATTRS select HAVE_DMA_ATTRS
select HAVE_DMA_API_DEBUG select HAVE_DMA_API_DEBUG
...@@ -47,7 +47,7 @@ config SPARC64 ...@@ -47,7 +47,7 @@ config SPARC64
select RTC_DRV_BQ4802 select RTC_DRV_BQ4802
select RTC_DRV_SUN4V select RTC_DRV_SUN4V
select RTC_DRV_STARFIRE select RTC_DRV_STARFIRE
select HAVE_PERF_COUNTERS select HAVE_PERF_EVENTS
config ARCH_DEFCONFIG config ARCH_DEFCONFIG
string string
......
#ifndef __ASM_SPARC_PERF_COUNTER_H
#define __ASM_SPARC_PERF_COUNTER_H
extern void set_perf_counter_pending(void);
#define PERF_COUNTER_INDEX_OFFSET 0
#ifdef CONFIG_PERF_COUNTERS
extern void init_hw_perf_counters(void);
#else
static inline void init_hw_perf_counters(void) { }
#endif
#endif
#ifndef __ASM_SPARC_PERF_EVENT_H
#define __ASM_SPARC_PERF_EVENT_H
extern void set_perf_event_pending(void);
#define PERF_EVENT_INDEX_OFFSET 0
#ifdef CONFIG_PERF_EVENTS
extern void init_hw_perf_events(void);
#else
static inline void init_hw_perf_events(void) { }
#endif
#endif
...@@ -395,7 +395,7 @@ ...@@ -395,7 +395,7 @@
#define __NR_preadv 324 #define __NR_preadv 324
#define __NR_pwritev 325 #define __NR_pwritev 325
#define __NR_rt_tgsigqueueinfo 326 #define __NR_rt_tgsigqueueinfo 326
#define __NR_perf_counter_open 327 #define __NR_perf_event_open 327
#define NR_SYSCALLS 328 #define NR_SYSCALLS 328
......
...@@ -104,5 +104,5 @@ obj-$(CONFIG_AUDIT) += audit.o ...@@ -104,5 +104,5 @@ obj-$(CONFIG_AUDIT) += audit.o
audit--$(CONFIG_AUDIT) := compat_audit.o audit--$(CONFIG_AUDIT) := compat_audit.o
obj-$(CONFIG_COMPAT) += $(audit--y) obj-$(CONFIG_COMPAT) += $(audit--y)
pc--$(CONFIG_PERF_COUNTERS) := perf_counter.o pc--$(CONFIG_PERF_EVENTS) := perf_event.o
obj-$(CONFIG_SPARC64) += $(pc--y) obj-$(CONFIG_SPARC64) += $(pc--y)
...@@ -19,7 +19,7 @@ ...@@ -19,7 +19,7 @@
#include <linux/delay.h> #include <linux/delay.h>
#include <linux/smp.h> #include <linux/smp.h>
#include <asm/perf_counter.h> #include <asm/perf_event.h>
#include <asm/ptrace.h> #include <asm/ptrace.h>
#include <asm/local.h> #include <asm/local.h>
#include <asm/pcr.h> #include <asm/pcr.h>
...@@ -265,7 +265,7 @@ int __init nmi_init(void) ...@@ -265,7 +265,7 @@ int __init nmi_init(void)
} }
} }
if (!err) if (!err)
init_hw_perf_counters(); init_hw_perf_events();
return err; return err;
} }
......
...@@ -7,7 +7,7 @@ ...@@ -7,7 +7,7 @@
#include <linux/init.h> #include <linux/init.h>
#include <linux/irq.h> #include <linux/irq.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <asm/pil.h> #include <asm/pil.h>
#include <asm/pcr.h> #include <asm/pcr.h>
...@@ -15,7 +15,7 @@ ...@@ -15,7 +15,7 @@
/* This code is shared between various users of the performance /* This code is shared between various users of the performance
* counters. Users will be oprofile, pseudo-NMI watchdog, and the * counters. Users will be oprofile, pseudo-NMI watchdog, and the
* perf_counter support layer. * perf_event support layer.
*/ */
#define PCR_SUN4U_ENABLE (PCR_PIC_PRIV | PCR_STRACE | PCR_UTRACE) #define PCR_SUN4U_ENABLE (PCR_PIC_PRIV | PCR_STRACE | PCR_UTRACE)
...@@ -42,14 +42,14 @@ void deferred_pcr_work_irq(int irq, struct pt_regs *regs) ...@@ -42,14 +42,14 @@ void deferred_pcr_work_irq(int irq, struct pt_regs *regs)
old_regs = set_irq_regs(regs); old_regs = set_irq_regs(regs);
irq_enter(); irq_enter();
#ifdef CONFIG_PERF_COUNTERS #ifdef CONFIG_PERF_EVENTS
perf_counter_do_pending(); perf_event_do_pending();
#endif #endif
irq_exit(); irq_exit();
set_irq_regs(old_regs); set_irq_regs(old_regs);
} }
void set_perf_counter_pending(void) void set_perf_event_pending(void)
{ {
set_softint(1 << PIL_DEFERRED_PCR_WORK); set_softint(1 << PIL_DEFERRED_PCR_WORK);
} }
......
...@@ -82,5 +82,5 @@ sys_call_table: ...@@ -82,5 +82,5 @@ sys_call_table:
/*310*/ .long sys_utimensat, sys_signalfd, sys_timerfd_create, sys_eventfd, sys_fallocate /*310*/ .long sys_utimensat, sys_signalfd, sys_timerfd_create, sys_eventfd, sys_fallocate
/*315*/ .long sys_timerfd_settime, sys_timerfd_gettime, sys_signalfd4, sys_eventfd2, sys_epoll_create1 /*315*/ .long sys_timerfd_settime, sys_timerfd_gettime, sys_signalfd4, sys_eventfd2, sys_epoll_create1
/*320*/ .long sys_dup3, sys_pipe2, sys_inotify_init1, sys_accept4, sys_preadv /*320*/ .long sys_dup3, sys_pipe2, sys_inotify_init1, sys_accept4, sys_preadv
/*325*/ .long sys_pwritev, sys_rt_tgsigqueueinfo, sys_perf_counter_open /*325*/ .long sys_pwritev, sys_rt_tgsigqueueinfo, sys_perf_event_open
...@@ -83,7 +83,7 @@ sys_call_table32: ...@@ -83,7 +83,7 @@ sys_call_table32:
/*310*/ .word compat_sys_utimensat, compat_sys_signalfd, sys_timerfd_create, sys_eventfd, compat_sys_fallocate /*310*/ .word compat_sys_utimensat, compat_sys_signalfd, sys_timerfd_create, sys_eventfd, compat_sys_fallocate
.word compat_sys_timerfd_settime, compat_sys_timerfd_gettime, compat_sys_signalfd4, sys_eventfd2, sys_epoll_create1 .word compat_sys_timerfd_settime, compat_sys_timerfd_gettime, compat_sys_signalfd4, sys_eventfd2, sys_epoll_create1
/*320*/ .word sys_dup3, sys_pipe2, sys_inotify_init1, sys_accept4, compat_sys_preadv /*320*/ .word sys_dup3, sys_pipe2, sys_inotify_init1, sys_accept4, compat_sys_preadv
.word compat_sys_pwritev, compat_sys_rt_tgsigqueueinfo, sys_perf_counter_open .word compat_sys_pwritev, compat_sys_rt_tgsigqueueinfo, sys_perf_event_open
#endif /* CONFIG_COMPAT */ #endif /* CONFIG_COMPAT */
...@@ -158,4 +158,4 @@ sys_call_table: ...@@ -158,4 +158,4 @@ sys_call_table:
/*310*/ .word sys_utimensat, sys_signalfd, sys_timerfd_create, sys_eventfd, sys_fallocate /*310*/ .word sys_utimensat, sys_signalfd, sys_timerfd_create, sys_eventfd, sys_fallocate
.word sys_timerfd_settime, sys_timerfd_gettime, sys_signalfd4, sys_eventfd2, sys_epoll_create1 .word sys_timerfd_settime, sys_timerfd_gettime, sys_signalfd4, sys_eventfd2, sys_epoll_create1
/*320*/ .word sys_dup3, sys_pipe2, sys_inotify_init1, sys_accept4, sys_preadv /*320*/ .word sys_dup3, sys_pipe2, sys_inotify_init1, sys_accept4, sys_preadv
.word sys_pwritev, sys_rt_tgsigqueueinfo, sys_perf_counter_open .word sys_pwritev, sys_rt_tgsigqueueinfo, sys_perf_event_open
...@@ -24,7 +24,7 @@ config X86 ...@@ -24,7 +24,7 @@ config X86
select HAVE_UNSTABLE_SCHED_CLOCK select HAVE_UNSTABLE_SCHED_CLOCK
select HAVE_IDE select HAVE_IDE
select HAVE_OPROFILE select HAVE_OPROFILE
select HAVE_PERF_COUNTERS if (!M386 && !M486) select HAVE_PERF_EVENTS if (!M386 && !M486)
select HAVE_IOREMAP_PROT select HAVE_IOREMAP_PROT
select HAVE_KPROBES select HAVE_KPROBES
select ARCH_WANT_OPTIONAL_GPIOLIB select ARCH_WANT_OPTIONAL_GPIOLIB
......
...@@ -831,5 +831,5 @@ ia32_sys_call_table: ...@@ -831,5 +831,5 @@ ia32_sys_call_table:
.quad compat_sys_preadv .quad compat_sys_preadv
.quad compat_sys_pwritev .quad compat_sys_pwritev
.quad compat_sys_rt_tgsigqueueinfo /* 335 */ .quad compat_sys_rt_tgsigqueueinfo /* 335 */
.quad sys_perf_counter_open .quad sys_perf_event_open
ia32_syscall_end: ia32_syscall_end:
...@@ -49,7 +49,7 @@ BUILD_INTERRUPT(apic_timer_interrupt,LOCAL_TIMER_VECTOR) ...@@ -49,7 +49,7 @@ BUILD_INTERRUPT(apic_timer_interrupt,LOCAL_TIMER_VECTOR)
BUILD_INTERRUPT(error_interrupt,ERROR_APIC_VECTOR) BUILD_INTERRUPT(error_interrupt,ERROR_APIC_VECTOR)
BUILD_INTERRUPT(spurious_interrupt,SPURIOUS_APIC_VECTOR) BUILD_INTERRUPT(spurious_interrupt,SPURIOUS_APIC_VECTOR)
#ifdef CONFIG_PERF_COUNTERS #ifdef CONFIG_PERF_EVENTS
BUILD_INTERRUPT(perf_pending_interrupt, LOCAL_PENDING_VECTOR) BUILD_INTERRUPT(perf_pending_interrupt, LOCAL_PENDING_VECTOR)
#endif #endif
......
#ifndef _ASM_X86_PERF_COUNTER_H #ifndef _ASM_X86_PERF_EVENT_H
#define _ASM_X86_PERF_COUNTER_H #define _ASM_X86_PERF_EVENT_H
/* /*
* Performance counter hw details: * Performance event hw details:
*/ */
#define X86_PMC_MAX_GENERIC 8 #define X86_PMC_MAX_GENERIC 8
...@@ -43,7 +43,7 @@ ...@@ -43,7 +43,7 @@
union cpuid10_eax { union cpuid10_eax {
struct { struct {
unsigned int version_id:8; unsigned int version_id:8;
unsigned int num_counters:8; unsigned int num_events:8;
unsigned int bit_width:8; unsigned int bit_width:8;
unsigned int mask_length:8; unsigned int mask_length:8;
} split; } split;
...@@ -52,7 +52,7 @@ union cpuid10_eax { ...@@ -52,7 +52,7 @@ union cpuid10_eax {
union cpuid10_edx { union cpuid10_edx {
struct { struct {
unsigned int num_counters_fixed:4; unsigned int num_events_fixed:4;
unsigned int reserved:28; unsigned int reserved:28;
} split; } split;
unsigned int full; unsigned int full;
...@@ -60,7 +60,7 @@ union cpuid10_edx { ...@@ -60,7 +60,7 @@ union cpuid10_edx {
/* /*
* Fixed-purpose performance counters: * Fixed-purpose performance events:
*/ */
/* /*
...@@ -87,22 +87,22 @@ union cpuid10_edx { ...@@ -87,22 +87,22 @@ union cpuid10_edx {
/* /*
* We model BTS tracing as another fixed-mode PMC. * We model BTS tracing as another fixed-mode PMC.
* *
* We choose a value in the middle of the fixed counter range, since lower * We choose a value in the middle of the fixed event range, since lower
* values are used by actual fixed counters and higher values are used * values are used by actual fixed events and higher values are used
* to indicate other overflow conditions in the PERF_GLOBAL_STATUS msr. * to indicate other overflow conditions in the PERF_GLOBAL_STATUS msr.
*/ */
#define X86_PMC_IDX_FIXED_BTS (X86_PMC_IDX_FIXED + 16) #define X86_PMC_IDX_FIXED_BTS (X86_PMC_IDX_FIXED + 16)
#ifdef CONFIG_PERF_COUNTERS #ifdef CONFIG_PERF_EVENTS
extern void init_hw_perf_counters(void); extern void init_hw_perf_events(void);
extern void perf_counters_lapic_init(void); extern void perf_events_lapic_init(void);
#define PERF_COUNTER_INDEX_OFFSET 0 #define PERF_EVENT_INDEX_OFFSET 0
#else #else
static inline void init_hw_perf_counters(void) { } static inline void init_hw_perf_events(void) { }
static inline void perf_counters_lapic_init(void) { } static inline void perf_events_lapic_init(void) { }
#endif #endif
#endif /* _ASM_X86_PERF_COUNTER_H */ #endif /* _ASM_X86_PERF_EVENT_H */
...@@ -341,7 +341,7 @@ ...@@ -341,7 +341,7 @@
#define __NR_preadv 333 #define __NR_preadv 333
#define __NR_pwritev 334 #define __NR_pwritev 334
#define __NR_rt_tgsigqueueinfo 335 #define __NR_rt_tgsigqueueinfo 335
#define __NR_perf_counter_open 336 #define __NR_perf_event_open 336
#ifdef __KERNEL__ #ifdef __KERNEL__
......
...@@ -659,8 +659,8 @@ __SYSCALL(__NR_preadv, sys_preadv) ...@@ -659,8 +659,8 @@ __SYSCALL(__NR_preadv, sys_preadv)
__SYSCALL(__NR_pwritev, sys_pwritev) __SYSCALL(__NR_pwritev, sys_pwritev)
#define __NR_rt_tgsigqueueinfo 297 #define __NR_rt_tgsigqueueinfo 297
__SYSCALL(__NR_rt_tgsigqueueinfo, sys_rt_tgsigqueueinfo) __SYSCALL(__NR_rt_tgsigqueueinfo, sys_rt_tgsigqueueinfo)
#define __NR_perf_counter_open 298 #define __NR_perf_event_open 298
__SYSCALL(__NR_perf_counter_open, sys_perf_counter_open) __SYSCALL(__NR_perf_event_open, sys_perf_event_open)
#ifndef __NO_STUBS #ifndef __NO_STUBS
#define __ARCH_WANT_OLD_READDIR #define __ARCH_WANT_OLD_READDIR
......
...@@ -14,7 +14,7 @@ ...@@ -14,7 +14,7 @@
* Mikael Pettersson : PM converted to driver model. * Mikael Pettersson : PM converted to driver model.
*/ */
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/kernel_stat.h> #include <linux/kernel_stat.h>
#include <linux/mc146818rtc.h> #include <linux/mc146818rtc.h>
#include <linux/acpi_pmtmr.h> #include <linux/acpi_pmtmr.h>
...@@ -35,7 +35,7 @@ ...@@ -35,7 +35,7 @@
#include <linux/smp.h> #include <linux/smp.h>
#include <linux/mm.h> #include <linux/mm.h>
#include <asm/perf_counter.h> #include <asm/perf_event.h>
#include <asm/x86_init.h> #include <asm/x86_init.h>
#include <asm/pgalloc.h> #include <asm/pgalloc.h>
#include <asm/atomic.h> #include <asm/atomic.h>
...@@ -1189,7 +1189,7 @@ void __cpuinit setup_local_APIC(void) ...@@ -1189,7 +1189,7 @@ void __cpuinit setup_local_APIC(void)
apic_write(APIC_ESR, 0); apic_write(APIC_ESR, 0);
} }
#endif #endif
perf_counters_lapic_init(); perf_events_lapic_init();
preempt_disable(); preempt_disable();
......
...@@ -27,7 +27,7 @@ obj-$(CONFIG_CPU_SUP_CENTAUR) += centaur.o ...@@ -27,7 +27,7 @@ obj-$(CONFIG_CPU_SUP_CENTAUR) += centaur.o
obj-$(CONFIG_CPU_SUP_TRANSMETA_32) += transmeta.o obj-$(CONFIG_CPU_SUP_TRANSMETA_32) += transmeta.o
obj-$(CONFIG_CPU_SUP_UMC_32) += umc.o obj-$(CONFIG_CPU_SUP_UMC_32) += umc.o
obj-$(CONFIG_PERF_COUNTERS) += perf_counter.o obj-$(CONFIG_PERF_EVENTS) += perf_event.o
obj-$(CONFIG_X86_MCE) += mcheck/ obj-$(CONFIG_X86_MCE) += mcheck/
obj-$(CONFIG_MTRR) += mtrr/ obj-$(CONFIG_MTRR) += mtrr/
......
...@@ -13,7 +13,7 @@ ...@@ -13,7 +13,7 @@
#include <linux/io.h> #include <linux/io.h>
#include <asm/stackprotector.h> #include <asm/stackprotector.h>
#include <asm/perf_counter.h> #include <asm/perf_event.h>
#include <asm/mmu_context.h> #include <asm/mmu_context.h>
#include <asm/hypervisor.h> #include <asm/hypervisor.h>
#include <asm/processor.h> #include <asm/processor.h>
...@@ -869,7 +869,7 @@ void __init identify_boot_cpu(void) ...@@ -869,7 +869,7 @@ void __init identify_boot_cpu(void)
#else #else
vgetcpu_set_mode(); vgetcpu_set_mode();
#endif #endif
init_hw_perf_counters(); init_hw_perf_events();
} }
void __cpuinit identify_secondary_cpu(struct cpuinfo_x86 *c) void __cpuinit identify_secondary_cpu(struct cpuinfo_x86 *c)
......
...@@ -20,7 +20,7 @@ ...@@ -20,7 +20,7 @@
#include <linux/kprobes.h> #include <linux/kprobes.h>
#include <asm/apic.h> #include <asm/apic.h>
#include <asm/perf_counter.h> #include <asm/perf_event.h>
struct nmi_watchdog_ctlblk { struct nmi_watchdog_ctlblk {
unsigned int cccr_msr; unsigned int cccr_msr;
......
...@@ -1021,7 +1021,7 @@ apicinterrupt ERROR_APIC_VECTOR \ ...@@ -1021,7 +1021,7 @@ apicinterrupt ERROR_APIC_VECTOR \
apicinterrupt SPURIOUS_APIC_VECTOR \ apicinterrupt SPURIOUS_APIC_VECTOR \
spurious_interrupt smp_spurious_interrupt spurious_interrupt smp_spurious_interrupt
#ifdef CONFIG_PERF_COUNTERS #ifdef CONFIG_PERF_EVENTS
apicinterrupt LOCAL_PENDING_VECTOR \ apicinterrupt LOCAL_PENDING_VECTOR \
perf_pending_interrupt smp_perf_pending_interrupt perf_pending_interrupt smp_perf_pending_interrupt
#endif #endif
......
...@@ -208,7 +208,7 @@ static void __init apic_intr_init(void) ...@@ -208,7 +208,7 @@ static void __init apic_intr_init(void)
alloc_intr_gate(ERROR_APIC_VECTOR, error_interrupt); alloc_intr_gate(ERROR_APIC_VECTOR, error_interrupt);
/* Performance monitoring interrupts: */ /* Performance monitoring interrupts: */
# ifdef CONFIG_PERF_COUNTERS # ifdef CONFIG_PERF_EVENTS
alloc_intr_gate(LOCAL_PENDING_VECTOR, perf_pending_interrupt); alloc_intr_gate(LOCAL_PENDING_VECTOR, perf_pending_interrupt);
# endif # endif
......
...@@ -335,4 +335,4 @@ ENTRY(sys_call_table) ...@@ -335,4 +335,4 @@ ENTRY(sys_call_table)
.long sys_preadv .long sys_preadv
.long sys_pwritev .long sys_pwritev
.long sys_rt_tgsigqueueinfo /* 335 */ .long sys_rt_tgsigqueueinfo /* 335 */
.long sys_perf_counter_open .long sys_perf_event_open
...@@ -10,7 +10,7 @@ ...@@ -10,7 +10,7 @@
#include <linux/bootmem.h> /* max_low_pfn */ #include <linux/bootmem.h> /* max_low_pfn */
#include <linux/kprobes.h> /* __kprobes, ... */ #include <linux/kprobes.h> /* __kprobes, ... */
#include <linux/mmiotrace.h> /* kmmio_handler, ... */ #include <linux/mmiotrace.h> /* kmmio_handler, ... */
#include <linux/perf_counter.h> /* perf_swcounter_event */ #include <linux/perf_event.h> /* perf_sw_event */
#include <asm/traps.h> /* dotraplinkage, ... */ #include <asm/traps.h> /* dotraplinkage, ... */
#include <asm/pgalloc.h> /* pgd_*(), ... */ #include <asm/pgalloc.h> /* pgd_*(), ... */
...@@ -1017,7 +1017,7 @@ do_page_fault(struct pt_regs *regs, unsigned long error_code) ...@@ -1017,7 +1017,7 @@ do_page_fault(struct pt_regs *regs, unsigned long error_code)
if (unlikely(error_code & PF_RSVD)) if (unlikely(error_code & PF_RSVD))
pgtable_bad(regs, error_code, address); pgtable_bad(regs, error_code, address);
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address);
/* /*
* If we're in an interrupt, have no user context or are running * If we're in an interrupt, have no user context or are running
...@@ -1114,11 +1114,11 @@ do_page_fault(struct pt_regs *regs, unsigned long error_code) ...@@ -1114,11 +1114,11 @@ do_page_fault(struct pt_regs *regs, unsigned long error_code)
if (fault & VM_FAULT_MAJOR) { if (fault & VM_FAULT_MAJOR) {
tsk->maj_flt++; tsk->maj_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0,
regs, address); regs, address);
} else { } else {
tsk->min_flt++; tsk->min_flt++;
perf_swcounter_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0,
regs, address); regs, address);
} }
......
...@@ -234,11 +234,11 @@ static void arch_perfmon_setup_counters(void) ...@@ -234,11 +234,11 @@ static void arch_perfmon_setup_counters(void)
if (eax.split.version_id == 0 && current_cpu_data.x86 == 6 && if (eax.split.version_id == 0 && current_cpu_data.x86 == 6 &&
current_cpu_data.x86_model == 15) { current_cpu_data.x86_model == 15) {
eax.split.version_id = 2; eax.split.version_id = 2;
eax.split.num_counters = 2; eax.split.num_events = 2;
eax.split.bit_width = 40; eax.split.bit_width = 40;
} }
num_counters = eax.split.num_counters; num_counters = eax.split.num_events;
op_arch_perfmon_spec.num_counters = num_counters; op_arch_perfmon_spec.num_counters = num_counters;
op_arch_perfmon_spec.num_controls = num_counters; op_arch_perfmon_spec.num_controls = num_counters;
......
...@@ -13,7 +13,7 @@ ...@@ -13,7 +13,7 @@
#define OP_X86_MODEL_H #define OP_X86_MODEL_H
#include <asm/types.h> #include <asm/types.h>
#include <asm/perf_counter.h> #include <asm/perf_event.h>
struct op_msr { struct op_msr {
unsigned long addr; unsigned long addr;
......
...@@ -26,7 +26,7 @@ ...@@ -26,7 +26,7 @@
#include <linux/proc_fs.h> #include <linux/proc_fs.h>
#include <linux/nmi.h> #include <linux/nmi.h>
#include <linux/quotaops.h> #include <linux/quotaops.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/suspend.h> #include <linux/suspend.h>
...@@ -252,7 +252,7 @@ static void sysrq_handle_showregs(int key, struct tty_struct *tty) ...@@ -252,7 +252,7 @@ static void sysrq_handle_showregs(int key, struct tty_struct *tty)
struct pt_regs *regs = get_irq_regs(); struct pt_regs *regs = get_irq_regs();
if (regs) if (regs)
show_regs(regs); show_regs(regs);
perf_counter_print_debug(); perf_event_print_debug();
} }
static struct sysrq_key_op sysrq_showregs_op = { static struct sysrq_key_op sysrq_showregs_op = {
.handler = sysrq_handle_showregs, .handler = sysrq_handle_showregs,
......
...@@ -33,7 +33,7 @@ ...@@ -33,7 +33,7 @@
#include <linux/string.h> #include <linux/string.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/pagemap.h> #include <linux/pagemap.h>
#include <linux/perf_counter.h> #include <linux/perf_event.h>
#include <linux/highmem.h> #include <linux/highmem.h>
#include <linux/spinlock.h> #include <linux/spinlock.h>
#include <linux/key.h> #include <linux/key.h>
...@@ -923,7 +923,7 @@ void set_task_comm(struct task_struct *tsk, char *buf) ...@@ -923,7 +923,7 @@ void set_task_comm(struct task_struct *tsk, char *buf)
task_lock(tsk); task_lock(tsk);
strlcpy(tsk->comm, buf, sizeof(tsk->comm)); strlcpy(tsk->comm, buf, sizeof(tsk->comm));
task_unlock(tsk); task_unlock(tsk);
perf_counter_comm(tsk); perf_event_comm(tsk);
} }
int flush_old_exec(struct linux_binprm * bprm) int flush_old_exec(struct linux_binprm * bprm)
...@@ -997,7 +997,7 @@ int flush_old_exec(struct linux_binprm * bprm) ...@@ -997,7 +997,7 @@ int flush_old_exec(struct linux_binprm * bprm)
* security domain: * security domain:
*/ */
if (!get_dumpable(current->mm)) if (!get_dumpable(current->mm))
perf_counter_exit_task(current); perf_event_exit_task(current);
/* An exec changes our domain. We are no longer part of the thread /* An exec changes our domain. We are no longer part of the thread
group */ group */
......
...@@ -620,8 +620,8 @@ __SYSCALL(__NR_move_pages, sys_move_pages) ...@@ -620,8 +620,8 @@ __SYSCALL(__NR_move_pages, sys_move_pages)
#define __NR_rt_tgsigqueueinfo 240 #define __NR_rt_tgsigqueueinfo 240
__SYSCALL(__NR_rt_tgsigqueueinfo, sys_rt_tgsigqueueinfo) __SYSCALL(__NR_rt_tgsigqueueinfo, sys_rt_tgsigqueueinfo)
#define __NR_perf_counter_open 241 #define __NR_perf_event_open 241
__SYSCALL(__NR_perf_counter_open, sys_perf_counter_open) __SYSCALL(__NR_perf_event_open, sys_perf_event_open)
#undef __NR_syscalls #undef __NR_syscalls
#define __NR_syscalls 242 #define __NR_syscalls 242
......
...@@ -106,13 +106,13 @@ extern struct group_info init_groups; ...@@ -106,13 +106,13 @@ extern struct group_info init_groups;
extern struct cred init_cred; extern struct cred init_cred;
#ifdef CONFIG_PERF_COUNTERS #ifdef CONFIG_PERF_EVENTS
# define INIT_PERF_COUNTERS(tsk) \ # define INIT_PERF_EVENTS(tsk) \
.perf_counter_mutex = \ .perf_event_mutex = \
__MUTEX_INITIALIZER(tsk.perf_counter_mutex), \ __MUTEX_INITIALIZER(tsk.perf_event_mutex), \
.perf_counter_list = LIST_HEAD_INIT(tsk.perf_counter_list), .perf_event_list = LIST_HEAD_INIT(tsk.perf_event_list),
#else #else
# define INIT_PERF_COUNTERS(tsk) # define INIT_PERF_EVENTS(tsk)
#endif #endif
/* /*
...@@ -178,7 +178,7 @@ extern struct cred init_cred; ...@@ -178,7 +178,7 @@ extern struct cred init_cred;
}, \ }, \
.dirties = INIT_PROP_LOCAL_SINGLE(dirties), \ .dirties = INIT_PROP_LOCAL_SINGLE(dirties), \
INIT_IDS \ INIT_IDS \
INIT_PERF_COUNTERS(tsk) \ INIT_PERF_EVENTS(tsk) \
INIT_TRACE_IRQFLAGS \ INIT_TRACE_IRQFLAGS \
INIT_LOCKDEP \ INIT_LOCKDEP \
INIT_FTRACE_GRAPH \ INIT_FTRACE_GRAPH \
......
This diff is collapsed.
This diff is collapsed.
...@@ -85,7 +85,7 @@ ...@@ -85,7 +85,7 @@
#define PR_SET_TIMERSLACK 29 #define PR_SET_TIMERSLACK 29
#define PR_GET_TIMERSLACK 30 #define PR_GET_TIMERSLACK 30
#define PR_TASK_PERF_COUNTERS_DISABLE 31 #define PR_TASK_PERF_EVENTS_DISABLE 31
#define PR_TASK_PERF_COUNTERS_ENABLE 32 #define PR_TASK_PERF_EVENTS_ENABLE 32
#endif /* _LINUX_PRCTL_H */ #endif /* _LINUX_PRCTL_H */
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment