Commit a78c942d authored by Chris Metcalf's avatar Chris Metcalf

arch/tile: parameterize system PLs to support KVM port

While not a port to KVM (yet), this change modifies the kernel
to be able to build either at PL1 or at PL2 with a suitable
config switch.  Pushing up this change avoids handling branch
merge issues going forward with the KVM work.
Signed-off-by: default avatarChris Metcalf <cmetcalf@tilera.com>
parent bf65e440
...@@ -96,6 +96,7 @@ config HVC_TILE ...@@ -96,6 +96,7 @@ config HVC_TILE
config TILE config TILE
def_bool y def_bool y
select HAVE_KVM if !TILEGX
select GENERIC_FIND_FIRST_BIT select GENERIC_FIND_FIRST_BIT
select GENERIC_FIND_NEXT_BIT select GENERIC_FIND_NEXT_BIT
select USE_GENERIC_SMP_HELPERS select USE_GENERIC_SMP_HELPERS
...@@ -314,6 +315,15 @@ config HARDWALL ...@@ -314,6 +315,15 @@ config HARDWALL
bool "Hardwall support to allow access to user dynamic network" bool "Hardwall support to allow access to user dynamic network"
default y default y
config KERNEL_PL
int "Processor protection level for kernel"
range 1 2
default "1"
---help---
This setting determines the processor protection level the
kernel will be built to run at. Generally you should use
the default value here.
endmenu # Tilera-specific configuration endmenu # Tilera-specific configuration
menu "Bus options" menu "Bus options"
...@@ -354,3 +364,5 @@ source "security/Kconfig" ...@@ -354,3 +364,5 @@ source "security/Kconfig"
source "crypto/Kconfig" source "crypto/Kconfig"
source "lib/Kconfig" source "lib/Kconfig"
source "arch/tile/kvm/Kconfig"
...@@ -53,6 +53,8 @@ libs-y += $(LIBGCC_PATH) ...@@ -53,6 +53,8 @@ libs-y += $(LIBGCC_PATH)
# See arch/tile/Kbuild for content of core part of the kernel # See arch/tile/Kbuild for content of core part of the kernel
core-y += arch/tile/ core-y += arch/tile/
core-$(CONFIG_KVM) += arch/tile/kvm/
ifdef TILERA_ROOT ifdef TILERA_ROOT
INSTALL_PATH ?= $(TILERA_ROOT)/tile/boot INSTALL_PATH ?= $(TILERA_ROOT)/tile/boot
endif endif
......
...@@ -12,8 +12,93 @@ ...@@ -12,8 +12,93 @@
* more details. * more details.
*/ */
/*
* In addition to including the proper base SPR definition file, depending
* on machine architecture, this file defines several macros which allow
* kernel code to use protection-level dependent SPRs without worrying
* about which PL it's running at. In these macros, the PL that the SPR
* or interrupt number applies to is replaced by K.
*/
#if CONFIG_KERNEL_PL != 1 && CONFIG_KERNEL_PL != 2
#error CONFIG_KERNEL_PL must be 1 or 2
#endif
/* Concatenate 4 strings. */
#define __concat4(a, b, c, d) a ## b ## c ## d
#define _concat4(a, b, c, d) __concat4(a, b, c, d)
#ifdef __tilegx__ #ifdef __tilegx__
#include <arch/spr_def_64.h> #include <arch/spr_def_64.h>
/* TILE-Gx dependent, protection-level dependent SPRs. */
#define SPR_INTERRUPT_MASK_K \
_concat4(SPR_INTERRUPT_MASK_, CONFIG_KERNEL_PL,,)
#define SPR_INTERRUPT_MASK_SET_K \
_concat4(SPR_INTERRUPT_MASK_SET_, CONFIG_KERNEL_PL,,)
#define SPR_INTERRUPT_MASK_RESET_K \
_concat4(SPR_INTERRUPT_MASK_RESET_, CONFIG_KERNEL_PL,,)
#define SPR_INTERRUPT_VECTOR_BASE_K \
_concat4(SPR_INTERRUPT_VECTOR_BASE_, CONFIG_KERNEL_PL,,)
#define SPR_IPI_MASK_K \
_concat4(SPR_IPI_MASK_, CONFIG_KERNEL_PL,,)
#define SPR_IPI_MASK_RESET_K \
_concat4(SPR_IPI_MASK_RESET_, CONFIG_KERNEL_PL,,)
#define SPR_IPI_MASK_SET_K \
_concat4(SPR_IPI_MASK_SET_, CONFIG_KERNEL_PL,,)
#define SPR_IPI_EVENT_K \
_concat4(SPR_IPI_EVENT_, CONFIG_KERNEL_PL,,)
#define SPR_IPI_EVENT_RESET_K \
_concat4(SPR_IPI_EVENT_RESET_, CONFIG_KERNEL_PL,,)
#define SPR_IPI_MASK_SET_K \
_concat4(SPR_IPI_MASK_SET_, CONFIG_KERNEL_PL,,)
#define INT_IPI_K \
_concat4(INT_IPI_, CONFIG_KERNEL_PL,,)
#define SPR_SINGLE_STEP_CONTROL_K \
_concat4(SPR_SINGLE_STEP_CONTROL_, CONFIG_KERNEL_PL,,)
#define SPR_SINGLE_STEP_EN_K_K \
_concat4(SPR_SINGLE_STEP_EN_, CONFIG_KERNEL_PL, _, CONFIG_KERNEL_PL)
#define INT_SINGLE_STEP_K \
_concat4(INT_SINGLE_STEP_, CONFIG_KERNEL_PL,,)
#else #else
#include <arch/spr_def_32.h> #include <arch/spr_def_32.h>
/* TILEPro dependent, protection-level dependent SPRs. */
#define SPR_INTERRUPT_MASK_K_0 \
_concat4(SPR_INTERRUPT_MASK_, CONFIG_KERNEL_PL, _0,)
#define SPR_INTERRUPT_MASK_K_1 \
_concat4(SPR_INTERRUPT_MASK_, CONFIG_KERNEL_PL, _1,)
#define SPR_INTERRUPT_MASK_SET_K_0 \
_concat4(SPR_INTERRUPT_MASK_SET_, CONFIG_KERNEL_PL, _0,)
#define SPR_INTERRUPT_MASK_SET_K_1 \
_concat4(SPR_INTERRUPT_MASK_SET_, CONFIG_KERNEL_PL, _1,)
#define SPR_INTERRUPT_MASK_RESET_K_0 \
_concat4(SPR_INTERRUPT_MASK_RESET_, CONFIG_KERNEL_PL, _0,)
#define SPR_INTERRUPT_MASK_RESET_K_1 \
_concat4(SPR_INTERRUPT_MASK_RESET_, CONFIG_KERNEL_PL, _1,)
#endif #endif
/* Generic protection-level dependent SPRs. */
#define SPR_SYSTEM_SAVE_K_0 \
_concat4(SPR_SYSTEM_SAVE_, CONFIG_KERNEL_PL, _0,)
#define SPR_SYSTEM_SAVE_K_1 \
_concat4(SPR_SYSTEM_SAVE_, CONFIG_KERNEL_PL, _1,)
#define SPR_SYSTEM_SAVE_K_2 \
_concat4(SPR_SYSTEM_SAVE_, CONFIG_KERNEL_PL, _2,)
#define SPR_SYSTEM_SAVE_K_3 \
_concat4(SPR_SYSTEM_SAVE_, CONFIG_KERNEL_PL, _3,)
#define SPR_EX_CONTEXT_K_0 \
_concat4(SPR_EX_CONTEXT_, CONFIG_KERNEL_PL, _0,)
#define SPR_EX_CONTEXT_K_1 \
_concat4(SPR_EX_CONTEXT_, CONFIG_KERNEL_PL, _1,)
#define SPR_INTCTRL_K_STATUS \
_concat4(SPR_INTCTRL_, CONFIG_KERNEL_PL, _STATUS,)
#define INT_INTCTRL_K \
_concat4(INT_INTCTRL_, CONFIG_KERNEL_PL,,)
...@@ -56,58 +56,93 @@ ...@@ -56,58 +56,93 @@
#define SPR_EX_CONTEXT_1_1__ICS_SHIFT 2 #define SPR_EX_CONTEXT_1_1__ICS_SHIFT 2
#define SPR_EX_CONTEXT_1_1__ICS_RMASK 0x1 #define SPR_EX_CONTEXT_1_1__ICS_RMASK 0x1
#define SPR_EX_CONTEXT_1_1__ICS_MASK 0x4 #define SPR_EX_CONTEXT_1_1__ICS_MASK 0x4
#define SPR_EX_CONTEXT_2_0 0x4605
#define SPR_EX_CONTEXT_2_1 0x4606
#define SPR_EX_CONTEXT_2_1__PL_SHIFT 0
#define SPR_EX_CONTEXT_2_1__PL_RMASK 0x3
#define SPR_EX_CONTEXT_2_1__PL_MASK 0x3
#define SPR_EX_CONTEXT_2_1__ICS_SHIFT 2
#define SPR_EX_CONTEXT_2_1__ICS_RMASK 0x1
#define SPR_EX_CONTEXT_2_1__ICS_MASK 0x4
#define SPR_FAIL 0x4e09 #define SPR_FAIL 0x4e09
#define SPR_INTCTRL_0_STATUS 0x4a07 #define SPR_INTCTRL_0_STATUS 0x4a07
#define SPR_INTCTRL_1_STATUS 0x4807 #define SPR_INTCTRL_1_STATUS 0x4807
#define SPR_INTCTRL_2_STATUS 0x4607
#define SPR_INTERRUPT_CRITICAL_SECTION 0x4e0a #define SPR_INTERRUPT_CRITICAL_SECTION 0x4e0a
#define SPR_INTERRUPT_MASK_0_0 0x4a08 #define SPR_INTERRUPT_MASK_0_0 0x4a08
#define SPR_INTERRUPT_MASK_0_1 0x4a09 #define SPR_INTERRUPT_MASK_0_1 0x4a09
#define SPR_INTERRUPT_MASK_1_0 0x4809 #define SPR_INTERRUPT_MASK_1_0 0x4809
#define SPR_INTERRUPT_MASK_1_1 0x480a #define SPR_INTERRUPT_MASK_1_1 0x480a
#define SPR_INTERRUPT_MASK_2_0 0x4608
#define SPR_INTERRUPT_MASK_2_1 0x4609
#define SPR_INTERRUPT_MASK_RESET_0_0 0x4a0a #define SPR_INTERRUPT_MASK_RESET_0_0 0x4a0a
#define SPR_INTERRUPT_MASK_RESET_0_1 0x4a0b #define SPR_INTERRUPT_MASK_RESET_0_1 0x4a0b
#define SPR_INTERRUPT_MASK_RESET_1_0 0x480b #define SPR_INTERRUPT_MASK_RESET_1_0 0x480b
#define SPR_INTERRUPT_MASK_RESET_1_1 0x480c #define SPR_INTERRUPT_MASK_RESET_1_1 0x480c
#define SPR_INTERRUPT_MASK_RESET_2_0 0x460a
#define SPR_INTERRUPT_MASK_RESET_2_1 0x460b
#define SPR_INTERRUPT_MASK_SET_0_0 0x4a0c #define SPR_INTERRUPT_MASK_SET_0_0 0x4a0c
#define SPR_INTERRUPT_MASK_SET_0_1 0x4a0d #define SPR_INTERRUPT_MASK_SET_0_1 0x4a0d
#define SPR_INTERRUPT_MASK_SET_1_0 0x480d #define SPR_INTERRUPT_MASK_SET_1_0 0x480d
#define SPR_INTERRUPT_MASK_SET_1_1 0x480e #define SPR_INTERRUPT_MASK_SET_1_1 0x480e
#define SPR_INTERRUPT_MASK_SET_2_0 0x460c
#define SPR_INTERRUPT_MASK_SET_2_1 0x460d
#define SPR_MPL_DMA_CPL_SET_0 0x5800 #define SPR_MPL_DMA_CPL_SET_0 0x5800
#define SPR_MPL_DMA_CPL_SET_1 0x5801 #define SPR_MPL_DMA_CPL_SET_1 0x5801
#define SPR_MPL_DMA_CPL_SET_2 0x5802
#define SPR_MPL_DMA_NOTIFY_SET_0 0x3800 #define SPR_MPL_DMA_NOTIFY_SET_0 0x3800
#define SPR_MPL_DMA_NOTIFY_SET_1 0x3801 #define SPR_MPL_DMA_NOTIFY_SET_1 0x3801
#define SPR_MPL_DMA_NOTIFY_SET_2 0x3802
#define SPR_MPL_INTCTRL_0_SET_0 0x4a00 #define SPR_MPL_INTCTRL_0_SET_0 0x4a00
#define SPR_MPL_INTCTRL_0_SET_1 0x4a01 #define SPR_MPL_INTCTRL_0_SET_1 0x4a01
#define SPR_MPL_INTCTRL_0_SET_2 0x4a02
#define SPR_MPL_INTCTRL_1_SET_0 0x4800 #define SPR_MPL_INTCTRL_1_SET_0 0x4800
#define SPR_MPL_INTCTRL_1_SET_1 0x4801 #define SPR_MPL_INTCTRL_1_SET_1 0x4801
#define SPR_MPL_INTCTRL_1_SET_2 0x4802
#define SPR_MPL_INTCTRL_2_SET_0 0x4600
#define SPR_MPL_INTCTRL_2_SET_1 0x4601
#define SPR_MPL_INTCTRL_2_SET_2 0x4602
#define SPR_MPL_SN_ACCESS_SET_0 0x0800 #define SPR_MPL_SN_ACCESS_SET_0 0x0800
#define SPR_MPL_SN_ACCESS_SET_1 0x0801 #define SPR_MPL_SN_ACCESS_SET_1 0x0801
#define SPR_MPL_SN_ACCESS_SET_2 0x0802
#define SPR_MPL_SN_CPL_SET_0 0x5a00 #define SPR_MPL_SN_CPL_SET_0 0x5a00
#define SPR_MPL_SN_CPL_SET_1 0x5a01 #define SPR_MPL_SN_CPL_SET_1 0x5a01
#define SPR_MPL_SN_CPL_SET_2 0x5a02
#define SPR_MPL_SN_FIREWALL_SET_0 0x2c00 #define SPR_MPL_SN_FIREWALL_SET_0 0x2c00
#define SPR_MPL_SN_FIREWALL_SET_1 0x2c01 #define SPR_MPL_SN_FIREWALL_SET_1 0x2c01
#define SPR_MPL_SN_FIREWALL_SET_2 0x2c02
#define SPR_MPL_SN_NOTIFY_SET_0 0x2a00 #define SPR_MPL_SN_NOTIFY_SET_0 0x2a00
#define SPR_MPL_SN_NOTIFY_SET_1 0x2a01 #define SPR_MPL_SN_NOTIFY_SET_1 0x2a01
#define SPR_MPL_SN_NOTIFY_SET_2 0x2a02
#define SPR_MPL_UDN_ACCESS_SET_0 0x0c00 #define SPR_MPL_UDN_ACCESS_SET_0 0x0c00
#define SPR_MPL_UDN_ACCESS_SET_1 0x0c01 #define SPR_MPL_UDN_ACCESS_SET_1 0x0c01
#define SPR_MPL_UDN_ACCESS_SET_2 0x0c02
#define SPR_MPL_UDN_AVAIL_SET_0 0x4000 #define SPR_MPL_UDN_AVAIL_SET_0 0x4000
#define SPR_MPL_UDN_AVAIL_SET_1 0x4001 #define SPR_MPL_UDN_AVAIL_SET_1 0x4001
#define SPR_MPL_UDN_AVAIL_SET_2 0x4002
#define SPR_MPL_UDN_CA_SET_0 0x3c00 #define SPR_MPL_UDN_CA_SET_0 0x3c00
#define SPR_MPL_UDN_CA_SET_1 0x3c01 #define SPR_MPL_UDN_CA_SET_1 0x3c01
#define SPR_MPL_UDN_CA_SET_2 0x3c02
#define SPR_MPL_UDN_COMPLETE_SET_0 0x1400 #define SPR_MPL_UDN_COMPLETE_SET_0 0x1400
#define SPR_MPL_UDN_COMPLETE_SET_1 0x1401 #define SPR_MPL_UDN_COMPLETE_SET_1 0x1401
#define SPR_MPL_UDN_COMPLETE_SET_2 0x1402
#define SPR_MPL_UDN_FIREWALL_SET_0 0x3000 #define SPR_MPL_UDN_FIREWALL_SET_0 0x3000
#define SPR_MPL_UDN_FIREWALL_SET_1 0x3001 #define SPR_MPL_UDN_FIREWALL_SET_1 0x3001
#define SPR_MPL_UDN_FIREWALL_SET_2 0x3002
#define SPR_MPL_UDN_REFILL_SET_0 0x1000 #define SPR_MPL_UDN_REFILL_SET_0 0x1000
#define SPR_MPL_UDN_REFILL_SET_1 0x1001 #define SPR_MPL_UDN_REFILL_SET_1 0x1001
#define SPR_MPL_UDN_REFILL_SET_2 0x1002
#define SPR_MPL_UDN_TIMER_SET_0 0x3600 #define SPR_MPL_UDN_TIMER_SET_0 0x3600
#define SPR_MPL_UDN_TIMER_SET_1 0x3601 #define SPR_MPL_UDN_TIMER_SET_1 0x3601
#define SPR_MPL_UDN_TIMER_SET_2 0x3602
#define SPR_MPL_WORLD_ACCESS_SET_0 0x4e00 #define SPR_MPL_WORLD_ACCESS_SET_0 0x4e00
#define SPR_MPL_WORLD_ACCESS_SET_1 0x4e01 #define SPR_MPL_WORLD_ACCESS_SET_1 0x4e01
#define SPR_MPL_WORLD_ACCESS_SET_2 0x4e02
#define SPR_PASS 0x4e0b #define SPR_PASS 0x4e0b
#define SPR_PERF_COUNT_0 0x4205 #define SPR_PERF_COUNT_0 0x4205
#define SPR_PERF_COUNT_1 0x4206 #define SPR_PERF_COUNT_1 0x4206
#define SPR_PERF_COUNT_CTL 0x4207 #define SPR_PERF_COUNT_CTL 0x4207
#define SPR_PERF_COUNT_DN_CTL 0x4210
#define SPR_PERF_COUNT_STS 0x4208 #define SPR_PERF_COUNT_STS 0x4208
#define SPR_PROC_STATUS 0x4f00 #define SPR_PROC_STATUS 0x4f00
#define SPR_SIM_CONTROL 0x4e0c #define SPR_SIM_CONTROL 0x4e0c
...@@ -124,6 +159,10 @@ ...@@ -124,6 +159,10 @@
#define SPR_SYSTEM_SAVE_1_1 0x4901 #define SPR_SYSTEM_SAVE_1_1 0x4901
#define SPR_SYSTEM_SAVE_1_2 0x4902 #define SPR_SYSTEM_SAVE_1_2 0x4902
#define SPR_SYSTEM_SAVE_1_3 0x4903 #define SPR_SYSTEM_SAVE_1_3 0x4903
#define SPR_SYSTEM_SAVE_2_0 0x4700
#define SPR_SYSTEM_SAVE_2_1 0x4701
#define SPR_SYSTEM_SAVE_2_2 0x4702
#define SPR_SYSTEM_SAVE_2_3 0x4703
#define SPR_TILE_COORD 0x4c17 #define SPR_TILE_COORD 0x4c17
#define SPR_TILE_RTF_HWM 0x4e10 #define SPR_TILE_RTF_HWM 0x4e10
#define SPR_TILE_TIMER_CONTROL 0x3205 #define SPR_TILE_TIMER_CONTROL 0x3205
......
...@@ -47,53 +47,53 @@ ...@@ -47,53 +47,53 @@
int __n = (n); \ int __n = (n); \
int __mask = 1 << (__n & 0x1f); \ int __mask = 1 << (__n & 0x1f); \
if (__n < 32) \ if (__n < 32) \
__insn_mtspr(SPR_INTERRUPT_MASK_SET_1_0, __mask); \ __insn_mtspr(SPR_INTERRUPT_MASK_SET_K_0, __mask); \
else \ else \
__insn_mtspr(SPR_INTERRUPT_MASK_SET_1_1, __mask); \ __insn_mtspr(SPR_INTERRUPT_MASK_SET_K_1, __mask); \
} while (0) } while (0)
#define interrupt_mask_reset(n) do { \ #define interrupt_mask_reset(n) do { \
int __n = (n); \ int __n = (n); \
int __mask = 1 << (__n & 0x1f); \ int __mask = 1 << (__n & 0x1f); \
if (__n < 32) \ if (__n < 32) \
__insn_mtspr(SPR_INTERRUPT_MASK_RESET_1_0, __mask); \ __insn_mtspr(SPR_INTERRUPT_MASK_RESET_K_0, __mask); \
else \ else \
__insn_mtspr(SPR_INTERRUPT_MASK_RESET_1_1, __mask); \ __insn_mtspr(SPR_INTERRUPT_MASK_RESET_K_1, __mask); \
} while (0) } while (0)
#define interrupt_mask_check(n) ({ \ #define interrupt_mask_check(n) ({ \
int __n = (n); \ int __n = (n); \
(((__n < 32) ? \ (((__n < 32) ? \
__insn_mfspr(SPR_INTERRUPT_MASK_1_0) : \ __insn_mfspr(SPR_INTERRUPT_MASK_K_0) : \
__insn_mfspr(SPR_INTERRUPT_MASK_1_1)) \ __insn_mfspr(SPR_INTERRUPT_MASK_K_1)) \
>> (__n & 0x1f)) & 1; \ >> (__n & 0x1f)) & 1; \
}) })
#define interrupt_mask_set_mask(mask) do { \ #define interrupt_mask_set_mask(mask) do { \
unsigned long long __m = (mask); \ unsigned long long __m = (mask); \
__insn_mtspr(SPR_INTERRUPT_MASK_SET_1_0, (unsigned long)(__m)); \ __insn_mtspr(SPR_INTERRUPT_MASK_SET_K_0, (unsigned long)(__m)); \
__insn_mtspr(SPR_INTERRUPT_MASK_SET_1_1, (unsigned long)(__m>>32)); \ __insn_mtspr(SPR_INTERRUPT_MASK_SET_K_1, (unsigned long)(__m>>32)); \
} while (0) } while (0)
#define interrupt_mask_reset_mask(mask) do { \ #define interrupt_mask_reset_mask(mask) do { \
unsigned long long __m = (mask); \ unsigned long long __m = (mask); \
__insn_mtspr(SPR_INTERRUPT_MASK_RESET_1_0, (unsigned long)(__m)); \ __insn_mtspr(SPR_INTERRUPT_MASK_RESET_K_0, (unsigned long)(__m)); \
__insn_mtspr(SPR_INTERRUPT_MASK_RESET_1_1, (unsigned long)(__m>>32)); \ __insn_mtspr(SPR_INTERRUPT_MASK_RESET_K_1, (unsigned long)(__m>>32)); \
} while (0) } while (0)
#else #else
#define interrupt_mask_set(n) \ #define interrupt_mask_set(n) \
__insn_mtspr(SPR_INTERRUPT_MASK_SET_1, (1UL << (n))) __insn_mtspr(SPR_INTERRUPT_MASK_SET_K, (1UL << (n)))
#define interrupt_mask_reset(n) \ #define interrupt_mask_reset(n) \
__insn_mtspr(SPR_INTERRUPT_MASK_RESET_1, (1UL << (n))) __insn_mtspr(SPR_INTERRUPT_MASK_RESET_K, (1UL << (n)))
#define interrupt_mask_check(n) \ #define interrupt_mask_check(n) \
((__insn_mfspr(SPR_INTERRUPT_MASK_1) >> (n)) & 1) ((__insn_mfspr(SPR_INTERRUPT_MASK_K) >> (n)) & 1)
#define interrupt_mask_set_mask(mask) \ #define interrupt_mask_set_mask(mask) \
__insn_mtspr(SPR_INTERRUPT_MASK_SET_1, (mask)) __insn_mtspr(SPR_INTERRUPT_MASK_SET_K, (mask))
#define interrupt_mask_reset_mask(mask) \ #define interrupt_mask_reset_mask(mask) \
__insn_mtspr(SPR_INTERRUPT_MASK_RESET_1, (mask)) __insn_mtspr(SPR_INTERRUPT_MASK_RESET_K, (mask))
#endif #endif
/* /*
* The set of interrupts we want active if irqs are enabled. * The set of interrupts we want active if irqs are enabled.
* Note that in particular, the tile timer interrupt comes and goes * Note that in particular, the tile timer interrupt comes and goes
* from this set, since we have no other way to turn off the timer. * from this set, since we have no other way to turn off the timer.
* Likewise, INTCTRL_1 is removed and re-added during device * Likewise, INTCTRL_K is removed and re-added during device
* interrupts, as is the the hardwall UDN_FIREWALL interrupt. * interrupts, as is the the hardwall UDN_FIREWALL interrupt.
* We use a low bit (MEM_ERROR) as our sentinel value and make sure it * We use a low bit (MEM_ERROR) as our sentinel value and make sure it
* is always claimed as an "active interrupt" so we can query that bit * is always claimed as an "active interrupt" so we can query that bit
...@@ -168,14 +168,14 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask); ...@@ -168,14 +168,14 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask);
/* Return 0 or 1 to indicate whether interrupts are currently disabled. */ /* Return 0 or 1 to indicate whether interrupts are currently disabled. */
#define IRQS_DISABLED(tmp) \ #define IRQS_DISABLED(tmp) \
mfspr tmp, INTERRUPT_MASK_1; \ mfspr tmp, SPR_INTERRUPT_MASK_K; \
andi tmp, tmp, 1 andi tmp, tmp, 1
/* Load up a pointer to &interrupts_enabled_mask. */ /* Load up a pointer to &interrupts_enabled_mask. */
#define GET_INTERRUPTS_ENABLED_MASK_PTR(reg) \ #define GET_INTERRUPTS_ENABLED_MASK_PTR(reg) \
moveli reg, hw2_last(interrupts_enabled_mask); \ moveli reg, hw2_last(interrupts_enabled_mask); \
shl16insli reg, reg, hw1(interrupts_enabled_mask); \ shl16insli reg, reg, hw1(interrupts_enabled_mask); \
shl16insli reg, reg, hw0(interrupts_enabled_mask); \ shl16insli reg, reg, hw0(interrupts_enabled_mask); \
add reg, reg, tp add reg, reg, tp
/* Disable interrupts. */ /* Disable interrupts. */
...@@ -183,18 +183,18 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask); ...@@ -183,18 +183,18 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask);
moveli tmp0, hw2_last(LINUX_MASKABLE_INTERRUPTS); \ moveli tmp0, hw2_last(LINUX_MASKABLE_INTERRUPTS); \
shl16insli tmp0, tmp0, hw1(LINUX_MASKABLE_INTERRUPTS); \ shl16insli tmp0, tmp0, hw1(LINUX_MASKABLE_INTERRUPTS); \
shl16insli tmp0, tmp0, hw0(LINUX_MASKABLE_INTERRUPTS); \ shl16insli tmp0, tmp0, hw0(LINUX_MASKABLE_INTERRUPTS); \
mtspr INTERRUPT_MASK_SET_1, tmp0 mtspr SPR_INTERRUPT_MASK_SET_K, tmp0
/* Disable ALL synchronous interrupts (used by NMI entry). */ /* Disable ALL synchronous interrupts (used by NMI entry). */
#define IRQ_DISABLE_ALL(tmp) \ #define IRQ_DISABLE_ALL(tmp) \
movei tmp, -1; \ movei tmp, -1; \
mtspr INTERRUPT_MASK_SET_1, tmp mtspr SPR_INTERRUPT_MASK_SET_K, tmp
/* Enable interrupts. */ /* Enable interrupts. */
#define IRQ_ENABLE(tmp0, tmp1) \ #define IRQ_ENABLE(tmp0, tmp1) \
GET_INTERRUPTS_ENABLED_MASK_PTR(tmp0); \ GET_INTERRUPTS_ENABLED_MASK_PTR(tmp0); \
ld tmp0, tmp0; \ ld tmp0, tmp0; \
mtspr INTERRUPT_MASK_RESET_1, tmp0 mtspr SPR_INTERRUPT_MASK_RESET_K, tmp0
#else /* !__tilegx__ */ #else /* !__tilegx__ */
...@@ -208,14 +208,14 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask); ...@@ -208,14 +208,14 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask);
* (making the original code's write of the "high" mask word idempotent). * (making the original code's write of the "high" mask word idempotent).
*/ */
#define IRQS_DISABLED(tmp) \ #define IRQS_DISABLED(tmp) \
mfspr tmp, INTERRUPT_MASK_1_0; \ mfspr tmp, SPR_INTERRUPT_MASK_K_0; \
shri tmp, tmp, INT_MEM_ERROR; \ shri tmp, tmp, INT_MEM_ERROR; \
andi tmp, tmp, 1 andi tmp, tmp, 1
/* Load up a pointer to &interrupts_enabled_mask. */ /* Load up a pointer to &interrupts_enabled_mask. */
#define GET_INTERRUPTS_ENABLED_MASK_PTR(reg) \ #define GET_INTERRUPTS_ENABLED_MASK_PTR(reg) \
moveli reg, lo16(interrupts_enabled_mask); \ moveli reg, lo16(interrupts_enabled_mask); \
auli reg, reg, ha16(interrupts_enabled_mask);\ auli reg, reg, ha16(interrupts_enabled_mask); \
add reg, reg, tp add reg, reg, tp
/* Disable interrupts. */ /* Disable interrupts. */
...@@ -225,16 +225,16 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask); ...@@ -225,16 +225,16 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask);
moveli tmp1, lo16(LINUX_MASKABLE_INTERRUPTS) \ moveli tmp1, lo16(LINUX_MASKABLE_INTERRUPTS) \
}; \ }; \
{ \ { \
mtspr INTERRUPT_MASK_SET_1_0, tmp0; \ mtspr SPR_INTERRUPT_MASK_SET_K_0, tmp0; \
auli tmp1, tmp1, ha16(LINUX_MASKABLE_INTERRUPTS) \ auli tmp1, tmp1, ha16(LINUX_MASKABLE_INTERRUPTS) \
}; \ }; \
mtspr INTERRUPT_MASK_SET_1_1, tmp1 mtspr SPR_INTERRUPT_MASK_SET_K_1, tmp1
/* Disable ALL synchronous interrupts (used by NMI entry). */ /* Disable ALL synchronous interrupts (used by NMI entry). */
#define IRQ_DISABLE_ALL(tmp) \ #define IRQ_DISABLE_ALL(tmp) \
movei tmp, -1; \ movei tmp, -1; \
mtspr INTERRUPT_MASK_SET_1_0, tmp; \ mtspr SPR_INTERRUPT_MASK_SET_K_0, tmp; \
mtspr INTERRUPT_MASK_SET_1_1, tmp mtspr SPR_INTERRUPT_MASK_SET_K_1, tmp
/* Enable interrupts. */ /* Enable interrupts. */
#define IRQ_ENABLE(tmp0, tmp1) \ #define IRQ_ENABLE(tmp0, tmp1) \
...@@ -244,8 +244,8 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask); ...@@ -244,8 +244,8 @@ DECLARE_PER_CPU(unsigned long long, interrupts_enabled_mask);
addi tmp1, tmp0, 4 \ addi tmp1, tmp0, 4 \
}; \ }; \
lw tmp1, tmp1; \ lw tmp1, tmp1; \
mtspr INTERRUPT_MASK_RESET_1_0, tmp0; \ mtspr SPR_INTERRUPT_MASK_RESET_K_0, tmp0; \
mtspr INTERRUPT_MASK_RESET_1_1, tmp1 mtspr SPR_INTERRUPT_MASK_RESET_K_1, tmp1
#endif #endif
/* /*
......
...@@ -199,17 +199,17 @@ static inline __attribute_const__ int get_order(unsigned long size) ...@@ -199,17 +199,17 @@ static inline __attribute_const__ int get_order(unsigned long size)
* If you want more physical memory than this then see the CONFIG_HIGHMEM * If you want more physical memory than this then see the CONFIG_HIGHMEM
* option in the kernel configuration. * option in the kernel configuration.
* *
* The top two 16MB chunks in the table below (VIRT and HV) are * The top 16MB chunk in the table below is unavailable to Linux. Since
* unavailable to Linux. Since the kernel interrupt vectors must live * the kernel interrupt vectors must live at ether 0xfe000000 or 0xfd000000
* at 0xfd000000, we map all of the bottom of RAM at this address with * (depending on whether the kernel is at PL2 or Pl1), we map all of the
* a huge page table entry to minimize its ITLB footprint (as well as * bottom of RAM at this address with a huge page table entry to minimize
* at PAGE_OFFSET). The last architected requirement is that user * its ITLB footprint (as well as at PAGE_OFFSET). The last architected
* interrupt vectors live at 0xfc000000, so we make that range of * requirement is that user interrupt vectors live at 0xfc000000, so we
* memory available to user processes. The remaining regions are sized * make that range of memory available to user processes. The remaining
* as shown; after the first four addresses, we show "typical" values, * regions are sized as shown; the first four addresses use the PL 1
* since the actual addresses depend on kernel #defines. * values, and after that, we show "typical" values, since the actual
* addresses depend on kernel #defines.
* *
* MEM_VIRT_INTRPT 0xff000000
* MEM_HV_INTRPT 0xfe000000 * MEM_HV_INTRPT 0xfe000000
* MEM_SV_INTRPT (kernel code) 0xfd000000 * MEM_SV_INTRPT (kernel code) 0xfd000000
* MEM_USER_INTRPT (user vector) 0xfc000000 * MEM_USER_INTRPT (user vector) 0xfc000000
...@@ -221,9 +221,14 @@ static inline __attribute_const__ int get_order(unsigned long size) ...@@ -221,9 +221,14 @@ static inline __attribute_const__ int get_order(unsigned long size)
*/ */
#define MEM_USER_INTRPT _AC(0xfc000000, UL) #define MEM_USER_INTRPT _AC(0xfc000000, UL)
#if CONFIG_KERNEL_PL == 1
#define MEM_SV_INTRPT _AC(0xfd000000, UL) #define MEM_SV_INTRPT _AC(0xfd000000, UL)
#define MEM_HV_INTRPT _AC(0xfe000000, UL) #define MEM_HV_INTRPT _AC(0xfe000000, UL)
#define MEM_VIRT_INTRPT _AC(0xff000000, UL) #else
#define MEM_GUEST_INTRPT _AC(0xfd000000, UL)
#define MEM_SV_INTRPT _AC(0xfe000000, UL)
#define MEM_HV_INTRPT _AC(0xff000000, UL)
#endif
#define INTRPT_SIZE 0x4000 #define INTRPT_SIZE 0x4000
......
...@@ -328,18 +328,21 @@ extern int kdata_huge; ...@@ -328,18 +328,21 @@ extern int kdata_huge;
* Note that assembly code assumes that USER_PL is zero. * Note that assembly code assumes that USER_PL is zero.
*/ */
#define USER_PL 0 #define USER_PL 0
#define KERNEL_PL 1 #if CONFIG_KERNEL_PL == 2
#define GUEST_PL 1
#endif
#define KERNEL_PL CONFIG_KERNEL_PL
/* SYSTEM_SAVE_1_0 holds the current cpu number ORed with ksp0. */ /* SYSTEM_SAVE_K_0 holds the current cpu number ORed with ksp0. */
#define CPU_LOG_MASK_VALUE 12 #define CPU_LOG_MASK_VALUE 12
#define CPU_MASK_VALUE ((1 << CPU_LOG_MASK_VALUE) - 1) #define CPU_MASK_VALUE ((1 << CPU_LOG_MASK_VALUE) - 1)
#if CONFIG_NR_CPUS > CPU_MASK_VALUE #if CONFIG_NR_CPUS > CPU_MASK_VALUE
# error Too many cpus! # error Too many cpus!
#endif #endif
#define raw_smp_processor_id() \ #define raw_smp_processor_id() \
((int)__insn_mfspr(SPR_SYSTEM_SAVE_1_0) & CPU_MASK_VALUE) ((int)__insn_mfspr(SPR_SYSTEM_SAVE_K_0) & CPU_MASK_VALUE)
#define get_current_ksp0() \ #define get_current_ksp0() \
(__insn_mfspr(SPR_SYSTEM_SAVE_1_0) & ~CPU_MASK_VALUE) (__insn_mfspr(SPR_SYSTEM_SAVE_K_0) & ~CPU_MASK_VALUE)
#define next_current_ksp0(task) ({ \ #define next_current_ksp0(task) ({ \
unsigned long __ksp0 = task_ksp0(task); \ unsigned long __ksp0 = task_ksp0(task); \
int __cpu = raw_smp_processor_id(); \ int __cpu = raw_smp_processor_id(); \
......
...@@ -62,8 +62,8 @@ struct pt_regs { ...@@ -62,8 +62,8 @@ struct pt_regs {
pt_reg_t lr; /* aliases regs[TREG_LR] */ pt_reg_t lr; /* aliases regs[TREG_LR] */
/* Saved special registers. */ /* Saved special registers. */
pt_reg_t pc; /* stored in EX_CONTEXT_1_0 */ pt_reg_t pc; /* stored in EX_CONTEXT_K_0 */
pt_reg_t ex1; /* stored in EX_CONTEXT_1_1 (PL and ICS bit) */ pt_reg_t ex1; /* stored in EX_CONTEXT_K_1 (PL and ICS bit) */
pt_reg_t faultnum; /* fault number (INT_SWINT_1 for syscall) */ pt_reg_t faultnum; /* fault number (INT_SWINT_1 for syscall) */
pt_reg_t orig_r0; /* r0 at syscall entry, else zero */ pt_reg_t orig_r0; /* r0 at syscall entry, else zero */
pt_reg_t flags; /* flags (see below) */ pt_reg_t flags; /* flags (see below) */
......
...@@ -164,7 +164,7 @@ extern struct task_struct *_switch_to(struct task_struct *prev, ...@@ -164,7 +164,7 @@ extern struct task_struct *_switch_to(struct task_struct *prev,
/* Helper function for _switch_to(). */ /* Helper function for _switch_to(). */
extern struct task_struct *__switch_to(struct task_struct *prev, extern struct task_struct *__switch_to(struct task_struct *prev,
struct task_struct *next, struct task_struct *next,
unsigned long new_system_save_1_0); unsigned long new_system_save_k_0);
/* Address that switched-away from tasks are at. */ /* Address that switched-away from tasks are at. */
extern unsigned long get_switch_to_pc(void); extern unsigned long get_switch_to_pc(void);
......
...@@ -1003,37 +1003,37 @@ int hv_console_write(HV_VirtAddr bytes, int len); ...@@ -1003,37 +1003,37 @@ int hv_console_write(HV_VirtAddr bytes, int len);
* when these occur in a client's interrupt critical section, they must * when these occur in a client's interrupt critical section, they must
* be delivered through the downcall mechanism. * be delivered through the downcall mechanism.
* *
* A downcall is initially delivered to the client as an INTCTRL_1 * A downcall is initially delivered to the client as an INTCTRL_CL
* interrupt. Upon entry to the INTCTRL_1 vector, the client must * interrupt, where CL is the client's PL. Upon entry to the INTCTRL_CL
* immediately invoke the hv_downcall_dispatch service. This service * vector, the client must immediately invoke the hv_downcall_dispatch
* will not return; instead it will cause one of the client's actual * service. This service will not return; instead it will cause one of
* downcall-handling interrupt vectors to be entered. The EX_CONTEXT * the client's actual downcall-handling interrupt vectors to be entered.
* registers in the client will be set so that when the client irets, * The EX_CONTEXT registers in the client will be set so that when the
* it will return to the code which was interrupted by the INTCTRL_1 * client irets, it will return to the code which was interrupted by the
* interrupt. * INTCTRL_CL interrupt.
* *
* Under some circumstances, the firing of INTCTRL_1 can race with * Under some circumstances, the firing of INTCTRL_CL can race with
* the lowering of a device interrupt. In such a case, the * the lowering of a device interrupt. In such a case, the
* hv_downcall_dispatch service may issue an iret instruction instead * hv_downcall_dispatch service may issue an iret instruction instead
* of entering one of the client's actual downcall-handling interrupt * of entering one of the client's actual downcall-handling interrupt
* vectors. This will return execution to the location that was * vectors. This will return execution to the location that was
* interrupted by INTCTRL_1. * interrupted by INTCTRL_CL.
* *
* Any saving of registers should be done by the actual handling * Any saving of registers should be done by the actual handling
* vectors; no registers should be changed by the INTCTRL_1 handler. * vectors; no registers should be changed by the INTCTRL_CL handler.
* In particular, the client should not use a jal instruction to invoke * In particular, the client should not use a jal instruction to invoke
* the hv_downcall_dispatch service, as that would overwrite the client's * the hv_downcall_dispatch service, as that would overwrite the client's
* lr register. Note that the hv_downcall_dispatch service may overwrite * lr register. Note that the hv_downcall_dispatch service may overwrite
* one or more of the client's system save registers. * one or more of the client's system save registers.
* *
* The client must not modify the INTCTRL_1_STATUS SPR. The hypervisor * The client must not modify the INTCTRL_CL_STATUS SPR. The hypervisor
* will set this register to cause a downcall to happen, and will clear * will set this register to cause a downcall to happen, and will clear
* it when no further downcalls are pending. * it when no further downcalls are pending.
* *
* When a downcall vector is entered, the INTCTRL_1 interrupt will be * When a downcall vector is entered, the INTCTRL_CL interrupt will be
* masked. When the client is done processing a downcall, and is ready * masked. When the client is done processing a downcall, and is ready
* to accept another, it must unmask this interrupt; if more downcalls * to accept another, it must unmask this interrupt; if more downcalls
* are pending, this will cause the INTCTRL_1 vector to be reentered. * are pending, this will cause the INTCTRL_CL vector to be reentered.
* Currently the following interrupt vectors can be entered through a * Currently the following interrupt vectors can be entered through a
* downcall: * downcall:
* *
......
...@@ -15,7 +15,9 @@ ...@@ -15,7 +15,9 @@
#include <linux/linkage.h> #include <linux/linkage.h>
#include <linux/unistd.h> #include <linux/unistd.h>
#include <asm/irqflags.h> #include <asm/irqflags.h>
#include <asm/processor.h>
#include <arch/abi.h> #include <arch/abi.h>
#include <arch/spr_def.h>
#ifdef __tilegx__ #ifdef __tilegx__
#define bnzt bnezt #define bnzt bnezt
...@@ -80,7 +82,7 @@ STD_ENTRY(KBacktraceIterator_init_current) ...@@ -80,7 +82,7 @@ STD_ENTRY(KBacktraceIterator_init_current)
STD_ENTRY(cpu_idle_on_new_stack) STD_ENTRY(cpu_idle_on_new_stack)
{ {
move sp, r1 move sp, r1
mtspr SYSTEM_SAVE_1_0, r2 mtspr SPR_SYSTEM_SAVE_K_0, r2
} }
jal free_thread_info jal free_thread_info
j cpu_idle j cpu_idle
...@@ -102,15 +104,15 @@ STD_ENTRY(smp_nap) ...@@ -102,15 +104,15 @@ STD_ENTRY(smp_nap)
STD_ENTRY(_cpu_idle) STD_ENTRY(_cpu_idle)
{ {
lnk r0 lnk r0
movei r1, 1 movei r1, KERNEL_PL
} }
{ {
addli r0, r0, _cpu_idle_nap - . addli r0, r0, _cpu_idle_nap - .
mtspr INTERRUPT_CRITICAL_SECTION, r1 mtspr INTERRUPT_CRITICAL_SECTION, r1
} }
IRQ_ENABLE(r2, r3) /* unmask, but still with ICS set */ IRQ_ENABLE(r2, r3) /* unmask, but still with ICS set */
mtspr EX_CONTEXT_1_1, r1 /* PL1, ICS clear */ mtspr SPR_EX_CONTEXT_K_1, r1 /* Kernel PL, ICS clear */
mtspr EX_CONTEXT_1_0, r0 mtspr SPR_EX_CONTEXT_K_0, r0
iret iret
.global _cpu_idle_nap .global _cpu_idle_nap
_cpu_idle_nap: _cpu_idle_nap:
......
...@@ -23,6 +23,7 @@ ...@@ -23,6 +23,7 @@
#include <asm/asm-offsets.h> #include <asm/asm-offsets.h>
#include <hv/hypervisor.h> #include <hv/hypervisor.h>
#include <arch/chip.h> #include <arch/chip.h>
#include <arch/spr_def.h>
/* /*
* This module contains the entry code for kernel images. It performs the * This module contains the entry code for kernel images. It performs the
...@@ -76,7 +77,7 @@ ENTRY(_start) ...@@ -76,7 +77,7 @@ ENTRY(_start)
} }
1: 1:
/* Get our processor number and save it away in SAVE_1_0. */ /* Get our processor number and save it away in SAVE_K_0. */
jal hv_inquire_topology jal hv_inquire_topology
mulll_uu r4, r1, r2 /* r1 == y, r2 == width */ mulll_uu r4, r1, r2 /* r1 == y, r2 == width */
add r4, r4, r0 /* r0 == x, so r4 == cpu == y*width + x */ add r4, r4, r0 /* r0 == x, so r4 == cpu == y*width + x */
...@@ -124,7 +125,7 @@ ENTRY(_start) ...@@ -124,7 +125,7 @@ ENTRY(_start)
lw r0, r0 lw r0, r0
lw sp, r1 lw sp, r1
or r4, sp, r4 or r4, sp, r4
mtspr SYSTEM_SAVE_1_0, r4 /* save ksp0 + cpu */ mtspr SPR_SYSTEM_SAVE_K_0, r4 /* save ksp0 + cpu */
addi sp, sp, -STACK_TOP_DELTA addi sp, sp, -STACK_TOP_DELTA
{ {
move lr, zero /* stop backtraces in the called function */ move lr, zero /* stop backtraces in the called function */
......
...@@ -32,8 +32,8 @@ ...@@ -32,8 +32,8 @@
# error "No support for kernel preemption currently" # error "No support for kernel preemption currently"
#endif #endif
#if INT_INTCTRL_1 < 32 || INT_INTCTRL_1 >= 48 #if INT_INTCTRL_K < 32 || INT_INTCTRL_K >= 48
# error INT_INTCTRL_1 coded to set high interrupt mask # error INT_INTCTRL_K coded to set high interrupt mask
#endif #endif
#define PTREGS_PTR(reg, ptreg) addli reg, sp, C_ABI_SAVE_AREA_SIZE + (ptreg) #define PTREGS_PTR(reg, ptreg) addli reg, sp, C_ABI_SAVE_AREA_SIZE + (ptreg)
...@@ -132,8 +132,8 @@ intvec_\vecname: ...@@ -132,8 +132,8 @@ intvec_\vecname:
/* Temporarily save a register so we have somewhere to work. */ /* Temporarily save a register so we have somewhere to work. */
mtspr SYSTEM_SAVE_1_1, r0 mtspr SPR_SYSTEM_SAVE_K_1, r0
mfspr r0, EX_CONTEXT_1_1 mfspr r0, SPR_EX_CONTEXT_K_1
/* The cmpxchg code clears sp to force us to reset it here on fault. */ /* The cmpxchg code clears sp to force us to reset it here on fault. */
{ {
...@@ -167,18 +167,18 @@ intvec_\vecname: ...@@ -167,18 +167,18 @@ intvec_\vecname:
* The page_fault handler may be downcalled directly by the * The page_fault handler may be downcalled directly by the
* hypervisor even when Linux is running and has ICS set. * hypervisor even when Linux is running and has ICS set.
* *
* In this case the contents of EX_CONTEXT_1_1 reflect the * In this case the contents of EX_CONTEXT_K_1 reflect the
* previous fault and can't be relied on to choose whether or * previous fault and can't be relied on to choose whether or
* not to reinitialize the stack pointer. So we add a test * not to reinitialize the stack pointer. So we add a test
* to see whether SYSTEM_SAVE_1_2 has the high bit set, * to see whether SYSTEM_SAVE_K_2 has the high bit set,
* and if so we don't reinitialize sp, since we must be coming * and if so we don't reinitialize sp, since we must be coming
* from Linux. (In fact the precise case is !(val & ~1), * from Linux. (In fact the precise case is !(val & ~1),
* but any Linux PC has to have the high bit set.) * but any Linux PC has to have the high bit set.)
* *
* Note that the hypervisor *always* sets SYSTEM_SAVE_1_2 for * Note that the hypervisor *always* sets SYSTEM_SAVE_K_2 for
* any path that turns into a downcall to one of our TLB handlers. * any path that turns into a downcall to one of our TLB handlers.
*/ */
mfspr r0, SYSTEM_SAVE_1_2 mfspr r0, SPR_SYSTEM_SAVE_K_2
{ {
blz r0, 0f /* high bit in S_S_1_2 is for a PC to use */ blz r0, 0f /* high bit in S_S_1_2 is for a PC to use */
move r0, sp move r0, sp
...@@ -187,12 +187,12 @@ intvec_\vecname: ...@@ -187,12 +187,12 @@ intvec_\vecname:
2: 2:
/* /*
* SYSTEM_SAVE_1_0 holds the cpu number in the low bits, and * SYSTEM_SAVE_K_0 holds the cpu number in the low bits, and
* the current stack top in the higher bits. So we recover * the current stack top in the higher bits. So we recover
* our stack top by just masking off the low bits, then * our stack top by just masking off the low bits, then
* point sp at the top aligned address on the actual stack page. * point sp at the top aligned address on the actual stack page.
*/ */
mfspr r0, SYSTEM_SAVE_1_0 mfspr r0, SPR_SYSTEM_SAVE_K_0
mm r0, r0, zero, LOG2_THREAD_SIZE, 31 mm r0, r0, zero, LOG2_THREAD_SIZE, 31
0: 0:
...@@ -254,7 +254,7 @@ intvec_\vecname: ...@@ -254,7 +254,7 @@ intvec_\vecname:
sw sp, r3 sw sp, r3
addli sp, sp, PTREGS_OFFSET_PC - PTREGS_OFFSET_REG(3) addli sp, sp, PTREGS_OFFSET_PC - PTREGS_OFFSET_REG(3)
} }
mfspr r0, EX_CONTEXT_1_0 mfspr r0, SPR_EX_CONTEXT_K_0
.ifc \processing,handle_syscall .ifc \processing,handle_syscall
/* /*
* Bump the saved PC by one bundle so that when we return, we won't * Bump the saved PC by one bundle so that when we return, we won't
...@@ -267,7 +267,7 @@ intvec_\vecname: ...@@ -267,7 +267,7 @@ intvec_\vecname:
sw sp, r0 sw sp, r0
addli sp, sp, PTREGS_OFFSET_EX1 - PTREGS_OFFSET_PC addli sp, sp, PTREGS_OFFSET_EX1 - PTREGS_OFFSET_PC
} }
mfspr r0, EX_CONTEXT_1_1 mfspr r0, SPR_EX_CONTEXT_K_1
{ {
sw sp, r0 sw sp, r0
addi sp, sp, PTREGS_OFFSET_FAULTNUM - PTREGS_OFFSET_EX1 addi sp, sp, PTREGS_OFFSET_FAULTNUM - PTREGS_OFFSET_EX1
...@@ -289,7 +289,7 @@ intvec_\vecname: ...@@ -289,7 +289,7 @@ intvec_\vecname:
.endif .endif
addli sp, sp, PTREGS_OFFSET_REG(0) - PTREGS_OFFSET_FAULTNUM addli sp, sp, PTREGS_OFFSET_REG(0) - PTREGS_OFFSET_FAULTNUM
} }
mfspr r0, SYSTEM_SAVE_1_1 /* Original r0 */ mfspr r0, SPR_SYSTEM_SAVE_K_1 /* Original r0 */
{ {
sw sp, r0 sw sp, r0
addi sp, sp, -PTREGS_OFFSET_REG(0) - 4 addi sp, sp, -PTREGS_OFFSET_REG(0) - 4
...@@ -309,12 +309,12 @@ intvec_\vecname: ...@@ -309,12 +309,12 @@ intvec_\vecname:
* See discussion below at "finish_interrupt_save". * See discussion below at "finish_interrupt_save".
*/ */
.ifc \c_routine, do_page_fault .ifc \c_routine, do_page_fault
mfspr r2, SYSTEM_SAVE_1_3 /* address of page fault */ mfspr r2, SPR_SYSTEM_SAVE_K_3 /* address of page fault */
mfspr r3, SYSTEM_SAVE_1_2 /* info about page fault */ mfspr r3, SPR_SYSTEM_SAVE_K_2 /* info about page fault */
.else .else
.ifc \vecnum, INT_DOUBLE_FAULT .ifc \vecnum, INT_DOUBLE_FAULT
{ {
mfspr r2, SYSTEM_SAVE_1_2 /* double fault info from HV */ mfspr r2, SPR_SYSTEM_SAVE_K_2 /* double fault info from HV */
movei r3, 0 movei r3, 0
} }
.else .else
...@@ -467,7 +467,7 @@ intvec_\vecname: ...@@ -467,7 +467,7 @@ intvec_\vecname:
/* Load tp with our per-cpu offset. */ /* Load tp with our per-cpu offset. */
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
{ {
mfspr r20, SYSTEM_SAVE_1_0 mfspr r20, SPR_SYSTEM_SAVE_K_0
moveli r21, lo16(__per_cpu_offset) moveli r21, lo16(__per_cpu_offset)
} }
{ {
...@@ -487,7 +487,7 @@ intvec_\vecname: ...@@ -487,7 +487,7 @@ intvec_\vecname:
* We load flags in r32 here so we can jump to .Lrestore_regs * We load flags in r32 here so we can jump to .Lrestore_regs
* directly after do_page_fault_ics() if necessary. * directly after do_page_fault_ics() if necessary.
*/ */
mfspr r32, EX_CONTEXT_1_1 mfspr r32, SPR_EX_CONTEXT_K_1
{ {
andi r32, r32, SPR_EX_CONTEXT_1_1__PL_MASK /* mask off ICS */ andi r32, r32, SPR_EX_CONTEXT_1_1__PL_MASK /* mask off ICS */
PTREGS_PTR(r21, PTREGS_OFFSET_FLAGS) PTREGS_PTR(r21, PTREGS_OFFSET_FLAGS)
...@@ -957,11 +957,11 @@ STD_ENTRY(interrupt_return) ...@@ -957,11 +957,11 @@ STD_ENTRY(interrupt_return)
pop_reg_zero r21, r3, sp, PTREGS_OFFSET_EX1 - PTREGS_OFFSET_PC pop_reg_zero r21, r3, sp, PTREGS_OFFSET_EX1 - PTREGS_OFFSET_PC
pop_reg_zero lr, r4, sp, PTREGS_OFFSET_REG(52) - PTREGS_OFFSET_EX1 pop_reg_zero lr, r4, sp, PTREGS_OFFSET_REG(52) - PTREGS_OFFSET_EX1
{ {
mtspr EX_CONTEXT_1_0, r21 mtspr SPR_EX_CONTEXT_K_0, r21
move r5, zero move r5, zero
} }
{ {
mtspr EX_CONTEXT_1_1, lr mtspr SPR_EX_CONTEXT_K_1, lr
andi lr, lr, SPR_EX_CONTEXT_1_1__PL_MASK /* mask off ICS */ andi lr, lr, SPR_EX_CONTEXT_1_1__PL_MASK /* mask off ICS */
} }
...@@ -1199,7 +1199,7 @@ STD_ENTRY(interrupt_return) ...@@ -1199,7 +1199,7 @@ STD_ENTRY(interrupt_return)
STD_ENDPROC(interrupt_return) STD_ENDPROC(interrupt_return)
/* /*
* This interrupt variant clears the INT_INTCTRL_1 interrupt mask bit * This interrupt variant clears the INT_INTCTRL_K interrupt mask bit
* before returning, so we can properly get more downcalls. * before returning, so we can properly get more downcalls.
*/ */
.pushsection .text.handle_interrupt_downcall,"ax" .pushsection .text.handle_interrupt_downcall,"ax"
...@@ -1208,11 +1208,11 @@ handle_interrupt_downcall: ...@@ -1208,11 +1208,11 @@ handle_interrupt_downcall:
check_single_stepping normal, .Ldispatch_downcall check_single_stepping normal, .Ldispatch_downcall
.Ldispatch_downcall: .Ldispatch_downcall:
/* Clear INTCTRL_1 from the set of interrupts we ever enable. */ /* Clear INTCTRL_K from the set of interrupts we ever enable. */
GET_INTERRUPTS_ENABLED_MASK_PTR(r30) GET_INTERRUPTS_ENABLED_MASK_PTR(r30)
{ {
addi r30, r30, 4 addi r30, r30, 4
movei r31, INT_MASK(INT_INTCTRL_1) movei r31, INT_MASK(INT_INTCTRL_K)
} }
{ {
lw r20, r30 lw r20, r30
...@@ -1227,7 +1227,7 @@ handle_interrupt_downcall: ...@@ -1227,7 +1227,7 @@ handle_interrupt_downcall:
} }
FEEDBACK_REENTER(handle_interrupt_downcall) FEEDBACK_REENTER(handle_interrupt_downcall)
/* Allow INTCTRL_1 to be enabled next time we enable interrupts. */ /* Allow INTCTRL_K to be enabled next time we enable interrupts. */
lw r20, r30 lw r20, r30
or r20, r20, r31 or r20, r20, r31
sw r30, r20 sw r30, r20
...@@ -1509,7 +1509,7 @@ handle_ill: ...@@ -1509,7 +1509,7 @@ handle_ill:
/* Various stub interrupt handlers and syscall handlers */ /* Various stub interrupt handlers and syscall handlers */
STD_ENTRY_LOCAL(_kernel_double_fault) STD_ENTRY_LOCAL(_kernel_double_fault)
mfspr r1, EX_CONTEXT_1_0 mfspr r1, SPR_EX_CONTEXT_K_0
move r2, lr move r2, lr
move r3, sp move r3, sp
move r4, r52 move r4, r52
...@@ -1518,7 +1518,7 @@ STD_ENTRY_LOCAL(_kernel_double_fault) ...@@ -1518,7 +1518,7 @@ STD_ENTRY_LOCAL(_kernel_double_fault)
STD_ENDPROC(_kernel_double_fault) STD_ENDPROC(_kernel_double_fault)
STD_ENTRY_LOCAL(bad_intr) STD_ENTRY_LOCAL(bad_intr)
mfspr r2, EX_CONTEXT_1_0 mfspr r2, SPR_EX_CONTEXT_K_0
panic "Unhandled interrupt %#x: PC %#lx" panic "Unhandled interrupt %#x: PC %#lx"
STD_ENDPROC(bad_intr) STD_ENDPROC(bad_intr)
...@@ -1560,7 +1560,7 @@ STD_ENTRY(_sys_clone) ...@@ -1560,7 +1560,7 @@ STD_ENTRY(_sys_clone)
* a page fault which would assume the stack was valid, it does * a page fault which would assume the stack was valid, it does
* save/restore the stack pointer and zero it out to make sure it gets reset. * save/restore the stack pointer and zero it out to make sure it gets reset.
* Since we always keep interrupts disabled, the hypervisor won't * Since we always keep interrupts disabled, the hypervisor won't
* clobber our EX_CONTEXT_1_x registers, so we don't save/restore them * clobber our EX_CONTEXT_K_x registers, so we don't save/restore them
* (other than to advance the PC on return). * (other than to advance the PC on return).
* *
* We have to manually validate the user vs kernel address range * We have to manually validate the user vs kernel address range
...@@ -1766,7 +1766,7 @@ ENTRY(sys_cmpxchg) ...@@ -1766,7 +1766,7 @@ ENTRY(sys_cmpxchg)
/* Do slow mtspr here so the following "mf" waits less. */ /* Do slow mtspr here so the following "mf" waits less. */
{ {
move sp, r27 move sp, r27
mtspr EX_CONTEXT_1_0, r28 mtspr SPR_EX_CONTEXT_K_0, r28
} }
mf mf
...@@ -1785,7 +1785,7 @@ ENTRY(sys_cmpxchg) ...@@ -1785,7 +1785,7 @@ ENTRY(sys_cmpxchg)
} }
{ {
move sp, r27 move sp, r27
mtspr EX_CONTEXT_1_0, r28 mtspr SPR_EX_CONTEXT_K_0, r28
} }
iret iret
...@@ -1813,7 +1813,7 @@ ENTRY(sys_cmpxchg) ...@@ -1813,7 +1813,7 @@ ENTRY(sys_cmpxchg)
#endif #endif
/* Issue the slow SPR here while the tns result is in flight. */ /* Issue the slow SPR here while the tns result is in flight. */
mfspr r28, EX_CONTEXT_1_0 mfspr r28, SPR_EX_CONTEXT_K_0
{ {
addi r28, r28, 8 /* return to the instruction after the swint1 */ addi r28, r28, 8 /* return to the instruction after the swint1 */
...@@ -1901,7 +1901,7 @@ ENTRY(sys_cmpxchg) ...@@ -1901,7 +1901,7 @@ ENTRY(sys_cmpxchg)
.Lcmpxchg64_mismatch: .Lcmpxchg64_mismatch:
{ {
move sp, r27 move sp, r27
mtspr EX_CONTEXT_1_0, r28 mtspr SPR_EX_CONTEXT_K_0, r28
} }
mf mf
{ {
...@@ -1982,8 +1982,13 @@ int_unalign: ...@@ -1982,8 +1982,13 @@ int_unalign:
int_hand INT_PERF_COUNT, PERF_COUNT, \ int_hand INT_PERF_COUNT, PERF_COUNT, \
op_handle_perf_interrupt, handle_nmi op_handle_perf_interrupt, handle_nmi
int_hand INT_INTCTRL_3, INTCTRL_3, bad_intr int_hand INT_INTCTRL_3, INTCTRL_3, bad_intr
#if CONFIG_KERNEL_PL == 2
dc_dispatch INT_INTCTRL_2, INTCTRL_2
int_hand INT_INTCTRL_1, INTCTRL_1, bad_intr
#else
int_hand INT_INTCTRL_2, INTCTRL_2, bad_intr int_hand INT_INTCTRL_2, INTCTRL_2, bad_intr
dc_dispatch INT_INTCTRL_1, INTCTRL_1 dc_dispatch INT_INTCTRL_1, INTCTRL_1
#endif
int_hand INT_INTCTRL_0, INTCTRL_0, bad_intr int_hand INT_INTCTRL_0, INTCTRL_0, bad_intr
int_hand INT_MESSAGE_RCV_DWNCL, MESSAGE_RCV_DWNCL, \ int_hand INT_MESSAGE_RCV_DWNCL, MESSAGE_RCV_DWNCL, \
hv_message_intr, handle_interrupt_downcall hv_message_intr, handle_interrupt_downcall
......
...@@ -61,9 +61,9 @@ static DEFINE_SPINLOCK(available_irqs_lock); ...@@ -61,9 +61,9 @@ static DEFINE_SPINLOCK(available_irqs_lock);
#if CHIP_HAS_IPI() #if CHIP_HAS_IPI()
/* Use SPRs to manipulate device interrupts. */ /* Use SPRs to manipulate device interrupts. */
#define mask_irqs(irq_mask) __insn_mtspr(SPR_IPI_MASK_SET_1, irq_mask) #define mask_irqs(irq_mask) __insn_mtspr(SPR_IPI_MASK_SET_K, irq_mask)
#define unmask_irqs(irq_mask) __insn_mtspr(SPR_IPI_MASK_RESET_1, irq_mask) #define unmask_irqs(irq_mask) __insn_mtspr(SPR_IPI_MASK_RESET_K, irq_mask)
#define clear_irqs(irq_mask) __insn_mtspr(SPR_IPI_EVENT_RESET_1, irq_mask) #define clear_irqs(irq_mask) __insn_mtspr(SPR_IPI_EVENT_RESET_K, irq_mask)
#else #else
/* Use HV to manipulate device interrupts. */ /* Use HV to manipulate device interrupts. */
#define mask_irqs(irq_mask) hv_disable_intr(irq_mask) #define mask_irqs(irq_mask) hv_disable_intr(irq_mask)
...@@ -89,16 +89,16 @@ void tile_dev_intr(struct pt_regs *regs, int intnum) ...@@ -89,16 +89,16 @@ void tile_dev_intr(struct pt_regs *regs, int intnum)
* masked by a previous interrupt. Then, mask out the ones * masked by a previous interrupt. Then, mask out the ones
* we're going to handle. * we're going to handle.
*/ */
unsigned long masked = __insn_mfspr(SPR_IPI_MASK_1); unsigned long masked = __insn_mfspr(SPR_IPI_MASK_K);
original_irqs = __insn_mfspr(SPR_IPI_EVENT_1) & ~masked; original_irqs = __insn_mfspr(SPR_IPI_EVENT_K) & ~masked;
__insn_mtspr(SPR_IPI_MASK_SET_1, original_irqs); __insn_mtspr(SPR_IPI_MASK_SET_K, original_irqs);
#else #else
/* /*
* Hypervisor performs the equivalent of the Gx code above and * Hypervisor performs the equivalent of the Gx code above and
* then puts the pending interrupt mask into a system save reg * then puts the pending interrupt mask into a system save reg
* for us to find. * for us to find.
*/ */
original_irqs = __insn_mfspr(SPR_SYSTEM_SAVE_1_3); original_irqs = __insn_mfspr(SPR_SYSTEM_SAVE_K_3);
#endif #endif
remaining_irqs = original_irqs; remaining_irqs = original_irqs;
...@@ -225,7 +225,7 @@ void __cpuinit setup_irq_regs(void) ...@@ -225,7 +225,7 @@ void __cpuinit setup_irq_regs(void)
/* Enable interrupt delivery. */ /* Enable interrupt delivery. */
unmask_irqs(~0UL); unmask_irqs(~0UL);
#if CHIP_HAS_IPI() #if CHIP_HAS_IPI()
raw_local_irq_unmask(INT_IPI_1); raw_local_irq_unmask(INT_IPI_K);
#endif #endif
} }
......
...@@ -34,7 +34,7 @@ void __cpuinit init_messaging(void) ...@@ -34,7 +34,7 @@ void __cpuinit init_messaging(void)
panic("hv_register_message_state: error %d", rc); panic("hv_register_message_state: error %d", rc);
/* Make sure downcall interrupts will be enabled. */ /* Make sure downcall interrupts will be enabled. */
raw_local_irq_unmask(INT_INTCTRL_1); raw_local_irq_unmask(INT_INTCTRL_K);
} }
void hv_message_intr(struct pt_regs *regs, int intnum) void hv_message_intr(struct pt_regs *regs, int intnum)
......
...@@ -305,15 +305,25 @@ int dump_task_regs(struct task_struct *tsk, elf_gregset_t *regs) ...@@ -305,15 +305,25 @@ int dump_task_regs(struct task_struct *tsk, elf_gregset_t *regs)
/* Allow user processes to access the DMA SPRs */ /* Allow user processes to access the DMA SPRs */
void grant_dma_mpls(void) void grant_dma_mpls(void)
{ {
#if CONFIG_KERNEL_PL == 2
__insn_mtspr(SPR_MPL_DMA_CPL_SET_1, 1);
__insn_mtspr(SPR_MPL_DMA_NOTIFY_SET_1, 1);
#else
__insn_mtspr(SPR_MPL_DMA_CPL_SET_0, 1); __insn_mtspr(SPR_MPL_DMA_CPL_SET_0, 1);
__insn_mtspr(SPR_MPL_DMA_NOTIFY_SET_0, 1); __insn_mtspr(SPR_MPL_DMA_NOTIFY_SET_0, 1);
#endif
} }
/* Forbid user processes from accessing the DMA SPRs */ /* Forbid user processes from accessing the DMA SPRs */
void restrict_dma_mpls(void) void restrict_dma_mpls(void)
{ {
#if CONFIG_KERNEL_PL == 2
__insn_mtspr(SPR_MPL_DMA_CPL_SET_2, 1);
__insn_mtspr(SPR_MPL_DMA_NOTIFY_SET_2, 1);
#else
__insn_mtspr(SPR_MPL_DMA_CPL_SET_1, 1); __insn_mtspr(SPR_MPL_DMA_CPL_SET_1, 1);
__insn_mtspr(SPR_MPL_DMA_NOTIFY_SET_1, 1); __insn_mtspr(SPR_MPL_DMA_NOTIFY_SET_1, 1);
#endif
} }
/* Pause the DMA engine, then save off its state registers. */ /* Pause the DMA engine, then save off its state registers. */
...@@ -524,7 +534,7 @@ struct task_struct *__sched _switch_to(struct task_struct *prev, ...@@ -524,7 +534,7 @@ struct task_struct *__sched _switch_to(struct task_struct *prev,
* Switch kernel SP, PC, and callee-saved registers. * Switch kernel SP, PC, and callee-saved registers.
* In the context of the new task, return the old task pointer * In the context of the new task, return the old task pointer
* (i.e. the task that actually called __switch_to). * (i.e. the task that actually called __switch_to).
* Pass the value to use for SYSTEM_SAVE_1_0 when we reset our sp. * Pass the value to use for SYSTEM_SAVE_K_0 when we reset our sp.
*/ */
return __switch_to(prev, next, next_current_ksp0(next)); return __switch_to(prev, next, next_current_ksp0(next));
} }
......
...@@ -85,7 +85,7 @@ STD_ENTRY_SECTION(__switch_to, .sched.text) ...@@ -85,7 +85,7 @@ STD_ENTRY_SECTION(__switch_to, .sched.text)
{ {
/* Update sp and ksp0 simultaneously to avoid backtracer warnings. */ /* Update sp and ksp0 simultaneously to avoid backtracer warnings. */
move sp, r13 move sp, r13
mtspr SYSTEM_SAVE_1_0, r2 mtspr SPR_SYSTEM_SAVE_K_0, r2
} }
FOR_EACH_CALLEE_SAVED_REG(LOAD_REG) FOR_EACH_CALLEE_SAVED_REG(LOAD_REG)
.L__switch_to_pc: .L__switch_to_pc:
......
...@@ -187,11 +187,11 @@ early_param("vmalloc", parse_vmalloc); ...@@ -187,11 +187,11 @@ early_param("vmalloc", parse_vmalloc);
#ifdef CONFIG_HIGHMEM #ifdef CONFIG_HIGHMEM
/* /*
* Determine for each controller where its lowmem is mapped and how * Determine for each controller where its lowmem is mapped and how much of
* much of it is mapped there. On controller zero, the first few * it is mapped there. On controller zero, the first few megabytes are
* megabytes are mapped at 0xfd000000 as code, so in principle we * already mapped in as code at MEM_SV_INTRPT, so in principle we could
* could start our data mappings higher up, but for now we don't * start our data mappings higher up, but for now we don't bother, to avoid
* bother, to avoid additional confusion. * additional confusion.
* *
* One question is whether, on systems with more than 768 Mb and * One question is whether, on systems with more than 768 Mb and
* controllers of different sizes, to map in a proportionate amount of * controllers of different sizes, to map in a proportionate amount of
...@@ -876,6 +876,9 @@ void __cpuinit setup_cpu(int boot) ...@@ -876,6 +876,9 @@ void __cpuinit setup_cpu(int boot)
#if CHIP_HAS_SN_PROC() #if CHIP_HAS_SN_PROC()
raw_local_irq_unmask(INT_SNITLB_MISS); raw_local_irq_unmask(INT_SNITLB_MISS);
#endif #endif
#ifdef __tilegx__
raw_local_irq_unmask(INT_SINGLE_STEP_K);
#endif
/* /*
* Allow user access to many generic SPRs, like the cycle * Allow user access to many generic SPRs, like the cycle
...@@ -893,11 +896,12 @@ void __cpuinit setup_cpu(int boot) ...@@ -893,11 +896,12 @@ void __cpuinit setup_cpu(int boot)
#endif #endif
/* /*
* Set the MPL for interrupt control 0 to user level. * Set the MPL for interrupt control 0 & 1 to the corresponding
* This includes access to the SYSTEM_SAVE and EX_CONTEXT SPRs, * values. This includes access to the SYSTEM_SAVE and EX_CONTEXT
* as well as the PL 0 interrupt mask. * SPRs, as well as the interrupt mask.
*/ */
__insn_mtspr(SPR_MPL_INTCTRL_0_SET_0, 1); __insn_mtspr(SPR_MPL_INTCTRL_0_SET_0, 1);
__insn_mtspr(SPR_MPL_INTCTRL_1_SET_1, 1);
/* Initialize IRQ support for this cpu. */ /* Initialize IRQ support for this cpu. */
setup_irq_regs(); setup_irq_regs();
...@@ -1033,7 +1037,7 @@ static void __init validate_va(void) ...@@ -1033,7 +1037,7 @@ static void __init validate_va(void)
* In addition, make sure we CAN'T use the end of memory, since * In addition, make sure we CAN'T use the end of memory, since
* we use the last chunk of each pgd for the pgd_list. * we use the last chunk of each pgd for the pgd_list.
*/ */
int i, fc_fd_ok = 0; int i, user_kernel_ok = 0;
unsigned long max_va = 0; unsigned long max_va = 0;
unsigned long list_va = unsigned long list_va =
((PGD_LIST_OFFSET / sizeof(pgd_t)) << PGDIR_SHIFT); ((PGD_LIST_OFFSET / sizeof(pgd_t)) << PGDIR_SHIFT);
...@@ -1044,13 +1048,13 @@ static void __init validate_va(void) ...@@ -1044,13 +1048,13 @@ static void __init validate_va(void)
break; break;
if (range.start <= MEM_USER_INTRPT && if (range.start <= MEM_USER_INTRPT &&
range.start + range.size >= MEM_HV_INTRPT) range.start + range.size >= MEM_HV_INTRPT)
fc_fd_ok = 1; user_kernel_ok = 1;
if (range.start == 0) if (range.start == 0)
max_va = range.size; max_va = range.size;
BUG_ON(range.start + range.size > list_va); BUG_ON(range.start + range.size > list_va);
} }
if (!fc_fd_ok) if (!user_kernel_ok)
early_panic("Hypervisor not configured for VAs 0xfc/0xfd\n"); early_panic("Hypervisor not configured for user/kernel VAs\n");
if (max_va == 0) if (max_va == 0)
early_panic("Hypervisor not configured for low VAs\n"); early_panic("Hypervisor not configured for low VAs\n");
if (max_va < KERNEL_HIGH_VADDR) if (max_va < KERNEL_HIGH_VADDR)
......
...@@ -212,7 +212,7 @@ void __init ipi_init(void) ...@@ -212,7 +212,7 @@ void __init ipi_init(void)
tile.x = cpu_x(cpu); tile.x = cpu_x(cpu);
tile.y = cpu_y(cpu); tile.y = cpu_y(cpu);
if (hv_get_ipi_pte(tile, 1, &pte) != 0) if (hv_get_ipi_pte(tile, KERNEL_PL, &pte) != 0)
panic("Failed to initialize IPI for cpu %d\n", cpu); panic("Failed to initialize IPI for cpu %d\n", cpu);
offset = hv_pte_get_pfn(pte) << PAGE_SHIFT; offset = hv_pte_get_pfn(pte) << PAGE_SHIFT;
......
...@@ -278,7 +278,7 @@ void __kprobes do_trap(struct pt_regs *regs, int fault_num, ...@@ -278,7 +278,7 @@ void __kprobes do_trap(struct pt_regs *regs, int fault_num,
case INT_DOUBLE_FAULT: case INT_DOUBLE_FAULT:
/* /*
* For double fault, "reason" is actually passed as * For double fault, "reason" is actually passed as
* SYSTEM_SAVE_1_2, the hypervisor's double-fault info, so * SYSTEM_SAVE_K_2, the hypervisor's double-fault info, so
* we can provide the original fault number rather than * we can provide the original fault number rather than
* the uninteresting "INT_DOUBLE_FAULT" so the user can * the uninteresting "INT_DOUBLE_FAULT" so the user can
* learn what actually struck while PL0 ICS was set. * learn what actually struck while PL0 ICS was set.
......
#
# KVM configuration
#
source "virt/kvm/Kconfig"
menuconfig VIRTUALIZATION
bool "Virtualization"
---help---
Say Y here to get to see options for using your Linux host to run
other operating systems inside virtual machines (guests).
This option alone does not add any kernel code.
If you say N, all options in this submenu will be skipped and
disabled.
if VIRTUALIZATION
config KVM
tristate "Kernel-based Virtual Machine (KVM) support"
depends on HAVE_KVM && MODULES && EXPERIMENTAL
select PREEMPT_NOTIFIERS
select ANON_INODES
---help---
Support hosting paravirtualized guest machines.
This module provides access to the hardware capabilities through
a character device node named /dev/kvm.
To compile this as a module, choose M here: the module
will be called kvm.
If unsure, say N.
source drivers/vhost/Kconfig
source drivers/virtio/Kconfig
endif # VIRTUALIZATION
...@@ -563,10 +563,10 @@ static int handle_page_fault(struct pt_regs *regs, ...@@ -563,10 +563,10 @@ static int handle_page_fault(struct pt_regs *regs,
/* /*
* When we take an ITLB or DTLB fault or access violation in the * When we take an ITLB or DTLB fault or access violation in the
* supervisor while the critical section bit is set, the hypervisor is * supervisor while the critical section bit is set, the hypervisor is
* reluctant to write new values into the EX_CONTEXT_1_x registers, * reluctant to write new values into the EX_CONTEXT_K_x registers,
* since that might indicate we have not yet squirreled the SPR * since that might indicate we have not yet squirreled the SPR
* contents away and can thus safely take a recursive interrupt. * contents away and can thus safely take a recursive interrupt.
* Accordingly, the hypervisor passes us the PC via SYSTEM_SAVE_1_2. * Accordingly, the hypervisor passes us the PC via SYSTEM_SAVE_K_2.
* *
* Note that this routine is called before homecache_tlb_defer_enter(), * Note that this routine is called before homecache_tlb_defer_enter(),
* which means that we can properly unlock any atomics that might * which means that we can properly unlock any atomics that might
...@@ -610,7 +610,7 @@ struct intvec_state do_page_fault_ics(struct pt_regs *regs, int fault_num, ...@@ -610,7 +610,7 @@ struct intvec_state do_page_fault_ics(struct pt_regs *regs, int fault_num,
* fault. We didn't set up a kernel stack on initial entry to * fault. We didn't set up a kernel stack on initial entry to
* sys_cmpxchg, but instead had one set up by the fault, which * sys_cmpxchg, but instead had one set up by the fault, which
* (because sys_cmpxchg never releases ICS) came to us via the * (because sys_cmpxchg never releases ICS) came to us via the
* SYSTEM_SAVE_1_2 mechanism, and thus EX_CONTEXT_1_[01] are * SYSTEM_SAVE_K_2 mechanism, and thus EX_CONTEXT_K_[01] are
* still referencing the original user code. We release the * still referencing the original user code. We release the
* atomic lock and rewrite pt_regs so that it appears that we * atomic lock and rewrite pt_regs so that it appears that we
* came from user-space directly, and after we finish the * came from user-space directly, and after we finish the
......
...@@ -1060,7 +1060,7 @@ void free_initmem(void) ...@@ -1060,7 +1060,7 @@ void free_initmem(void)
/* /*
* Free the pages mapped from 0xc0000000 that correspond to code * Free the pages mapped from 0xc0000000 that correspond to code
* pages from 0xfd000000 that we won't use again after init. * pages from MEM_SV_INTRPT that we won't use again after init.
*/ */
free_init_pages("unused kernel text", free_init_pages("unused kernel text",
(unsigned long)_sinittext - text_delta, (unsigned long)_sinittext - text_delta,
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment