Commit 03f511dd authored by Isaku Yamahata's avatar Isaku Yamahata Committed by Tony Luck

ia64/pv_ops: implement binary patching optimization for native.

implement binary patching optimization for pv_cpu_ops.
With this optimization, indirect call for pv_cpu_ops methods can be
converted into inline execution or direct call.
Signed-off-by: default avatarIsaku Yamahata <yamahata@valinux.co.jp>
Signed-off-by: default avatarTony Luck <tony.luck@intel.com>
parent bf7ab02f
...@@ -201,7 +201,11 @@ extern long ia64_cmpxchg_called_with_bad_pointer (void); ...@@ -201,7 +201,11 @@ extern long ia64_cmpxchg_called_with_bad_pointer (void);
#ifndef __ASSEMBLY__ #ifndef __ASSEMBLY__
#if defined(CONFIG_PARAVIRT) && defined(__KERNEL__) #if defined(CONFIG_PARAVIRT) && defined(__KERNEL__)
#define IA64_INTRINSIC_API(name) pv_cpu_ops.name #ifdef ASM_SUPPORTED
# define IA64_INTRINSIC_API(name) paravirt_ ## name
#else
# define IA64_INTRINSIC_API(name) pv_cpu_ops.name
#endif
#define IA64_INTRINSIC_MACRO(name) paravirt_ ## name #define IA64_INTRINSIC_MACRO(name) paravirt_ ## name
#else #else
#define IA64_INTRINSIC_API(name) ia64_native_ ## name #define IA64_INTRINSIC_API(name) ia64_native_ ## name
......
...@@ -118,6 +118,14 @@ struct pv_init_ops { ...@@ -118,6 +118,14 @@ struct pv_init_ops {
int (*arch_setup_nomca)(void); int (*arch_setup_nomca)(void);
void (*post_smp_prepare_boot_cpu)(void); void (*post_smp_prepare_boot_cpu)(void);
#ifdef ASM_SUPPORTED
unsigned long (*patch_bundle)(void *sbundle, void *ebundle,
unsigned long type);
unsigned long (*patch_inst)(unsigned long stag, unsigned long etag,
unsigned long type);
#endif
void (*patch_branch)(unsigned long tag, unsigned long type);
}; };
extern struct pv_init_ops pv_init_ops; extern struct pv_init_ops pv_init_ops;
......
This diff is collapsed.
...@@ -36,7 +36,8 @@ obj-$(CONFIG_PCI_MSI) += msi_ia64.o ...@@ -36,7 +36,8 @@ obj-$(CONFIG_PCI_MSI) += msi_ia64.o
mca_recovery-y += mca_drv.o mca_drv_asm.o mca_recovery-y += mca_drv.o mca_drv_asm.o
obj-$(CONFIG_IA64_MC_ERR_INJECT)+= err_inject.o obj-$(CONFIG_IA64_MC_ERR_INJECT)+= err_inject.o
obj-$(CONFIG_PARAVIRT) += paravirt.o paravirtentry.o obj-$(CONFIG_PARAVIRT) += paravirt.o paravirtentry.o \
paravirt_patch.o
obj-$(CONFIG_IA64_ESI) += esi.o obj-$(CONFIG_IA64_ESI) += esi.o
ifneq ($(CONFIG_IA64_ESI),) ifneq ($(CONFIG_IA64_ESI),)
......
This diff is collapsed.
...@@ -20,8 +20,11 @@ ...@@ -20,8 +20,11 @@
* *
*/ */
#include <linux/init.h>
#include <asm/asmmacro.h> #include <asm/asmmacro.h>
#include <asm/asm-offsets.h> #include <asm/asm-offsets.h>
#include <asm/paravirt_privop.h>
#include <asm/paravirt_patch.h>
#include "entry.h" #include "entry.h"
#define DATA8(sym, init_value) \ #define DATA8(sym, init_value) \
...@@ -32,32 +35,34 @@ ...@@ -32,32 +35,34 @@
data8 init_value ; \ data8 init_value ; \
.popsection .popsection
#define BRANCH(targ, reg, breg) \ #define BRANCH(targ, reg, breg, type) \
movl reg=targ ; \ PARAVIRT_PATCH_SITE_BR(PARAVIRT_PATCH_TYPE_BR_ ## type) ; \
;; \ ;; \
ld8 reg=[reg] ; \ movl reg=targ ; \
;; \ ;; \
mov breg=reg ; \ ld8 reg=[reg] ; \
;; \
mov breg=reg ; \
br.cond.sptk.many breg br.cond.sptk.many breg
#define BRANCH_PROC(sym, reg, breg) \ #define BRANCH_PROC(sym, reg, breg, type) \
DATA8(paravirt_ ## sym ## _targ, ia64_native_ ## sym) ; \ DATA8(paravirt_ ## sym ## _targ, ia64_native_ ## sym) ; \
GLOBAL_ENTRY(paravirt_ ## sym) ; \ GLOBAL_ENTRY(paravirt_ ## sym) ; \
BRANCH(paravirt_ ## sym ## _targ, reg, breg) ; \ BRANCH(paravirt_ ## sym ## _targ, reg, breg, type) ; \
END(paravirt_ ## sym) END(paravirt_ ## sym)
#define BRANCH_PROC_UNWINFO(sym, reg, breg) \ #define BRANCH_PROC_UNWINFO(sym, reg, breg, type) \
DATA8(paravirt_ ## sym ## _targ, ia64_native_ ## sym) ; \ DATA8(paravirt_ ## sym ## _targ, ia64_native_ ## sym) ; \
GLOBAL_ENTRY(paravirt_ ## sym) ; \ GLOBAL_ENTRY(paravirt_ ## sym) ; \
PT_REGS_UNWIND_INFO(0) ; \ PT_REGS_UNWIND_INFO(0) ; \
BRANCH(paravirt_ ## sym ## _targ, reg, breg) ; \ BRANCH(paravirt_ ## sym ## _targ, reg, breg, type) ; \
END(paravirt_ ## sym) END(paravirt_ ## sym)
BRANCH_PROC(switch_to, r22, b7) BRANCH_PROC(switch_to, r22, b7, SWITCH_TO)
BRANCH_PROC_UNWINFO(leave_syscall, r22, b7) BRANCH_PROC_UNWINFO(leave_syscall, r22, b7, LEAVE_SYSCALL)
BRANCH_PROC(work_processed_syscall, r2, b7) BRANCH_PROC(work_processed_syscall, r2, b7, WORK_PROCESSED_SYSCALL)
BRANCH_PROC_UNWINFO(leave_kernel, r22, b7) BRANCH_PROC_UNWINFO(leave_kernel, r22, b7, LEAVE_KERNEL)
#ifdef CONFIG_MODULES #ifdef CONFIG_MODULES
......
...@@ -52,6 +52,7 @@ ...@@ -52,6 +52,7 @@
#include <asm/meminit.h> #include <asm/meminit.h>
#include <asm/page.h> #include <asm/page.h>
#include <asm/paravirt.h> #include <asm/paravirt.h>
#include <asm/paravirt_patch.h>
#include <asm/patch.h> #include <asm/patch.h>
#include <asm/pgtable.h> #include <asm/pgtable.h>
#include <asm/processor.h> #include <asm/processor.h>
...@@ -537,6 +538,7 @@ setup_arch (char **cmdline_p) ...@@ -537,6 +538,7 @@ setup_arch (char **cmdline_p)
paravirt_arch_setup_early(); paravirt_arch_setup_early();
ia64_patch_vtop((u64) __start___vtop_patchlist, (u64) __end___vtop_patchlist); ia64_patch_vtop((u64) __start___vtop_patchlist, (u64) __end___vtop_patchlist);
paravirt_patch_apply();
*cmdline_p = __va(ia64_boot_param->command_line); *cmdline_p = __va(ia64_boot_param->command_line);
strlcpy(boot_command_line, *cmdline_p, COMMAND_LINE_SIZE); strlcpy(boot_command_line, *cmdline_p, COMMAND_LINE_SIZE);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment