308 lines
7.1 KiB
ArmAsm
308 lines
7.1 KiB
ArmAsm
|
/* SPDX-License-Identifier: GPL-2.0-only */
|
||
|
/*
|
||
|
* Hypervisor stub
|
||
|
*
|
||
|
* Copyright (C) 2012 ARM Ltd.
|
||
|
* Author: Marc Zyngier <marc.zyngier@arm.com>
|
||
|
*/
|
||
|
|
||
|
#include <linux/init.h>
|
||
|
#include <linux/linkage.h>
|
||
|
|
||
|
#include <asm/assembler.h>
|
||
|
#include <asm/el2_setup.h>
|
||
|
#include <asm/kvm_arm.h>
|
||
|
#include <asm/kvm_asm.h>
|
||
|
#include <asm/ptrace.h>
|
||
|
#include <asm/virt.h>
|
||
|
|
||
|
// Warning, hardcoded register allocation
|
||
|
// This will clobber x1 and x2, and expect x1 to contain
|
||
|
// the id register value as read from the HW
|
||
|
.macro __check_override idreg, fld, width, pass, fail
|
||
|
ubfx x1, x1, #\fld, #\width
|
||
|
cbz x1, \fail
|
||
|
|
||
|
adr_l x1, \idreg\()_override
|
||
|
ldr x2, [x1, FTR_OVR_VAL_OFFSET]
|
||
|
ldr x1, [x1, FTR_OVR_MASK_OFFSET]
|
||
|
ubfx x2, x2, #\fld, #\width
|
||
|
ubfx x1, x1, #\fld, #\width
|
||
|
cmp x1, xzr
|
||
|
and x2, x2, x1
|
||
|
csinv x2, x2, xzr, ne
|
||
|
cbnz x2, \pass
|
||
|
b \fail
|
||
|
.endm
|
||
|
|
||
|
.macro check_override idreg, fld, pass, fail
|
||
|
mrs x1, \idreg\()_el1
|
||
|
__check_override \idreg \fld 4 \pass \fail
|
||
|
.endm
|
||
|
|
||
|
.text
|
||
|
.pushsection .hyp.text, "ax"
|
||
|
|
||
|
.align 11
|
||
|
|
||
|
SYM_CODE_START(__hyp_stub_vectors)
|
||
|
ventry el2_sync_invalid // Synchronous EL2t
|
||
|
ventry el2_irq_invalid // IRQ EL2t
|
||
|
ventry el2_fiq_invalid // FIQ EL2t
|
||
|
ventry el2_error_invalid // Error EL2t
|
||
|
|
||
|
ventry elx_sync // Synchronous EL2h
|
||
|
ventry el2_irq_invalid // IRQ EL2h
|
||
|
ventry el2_fiq_invalid // FIQ EL2h
|
||
|
ventry el2_error_invalid // Error EL2h
|
||
|
|
||
|
ventry elx_sync // Synchronous 64-bit EL1
|
||
|
ventry el1_irq_invalid // IRQ 64-bit EL1
|
||
|
ventry el1_fiq_invalid // FIQ 64-bit EL1
|
||
|
ventry el1_error_invalid // Error 64-bit EL1
|
||
|
|
||
|
ventry el1_sync_invalid // Synchronous 32-bit EL1
|
||
|
ventry el1_irq_invalid // IRQ 32-bit EL1
|
||
|
ventry el1_fiq_invalid // FIQ 32-bit EL1
|
||
|
ventry el1_error_invalid // Error 32-bit EL1
|
||
|
SYM_CODE_END(__hyp_stub_vectors)
|
||
|
|
||
|
.align 11
|
||
|
|
||
|
SYM_CODE_START_LOCAL(elx_sync)
|
||
|
cmp x0, #HVC_SET_VECTORS
|
||
|
b.ne 1f
|
||
|
msr vbar_el2, x1
|
||
|
b 9f
|
||
|
|
||
|
1: cmp x0, #HVC_FINALISE_EL2
|
||
|
b.eq __finalise_el2
|
||
|
|
||
|
2: cmp x0, #HVC_SOFT_RESTART
|
||
|
b.ne 3f
|
||
|
mov x0, x2
|
||
|
mov x2, x4
|
||
|
mov x4, x1
|
||
|
mov x1, x3
|
||
|
br x4 // no return
|
||
|
|
||
|
3: cmp x0, #HVC_RESET_VECTORS
|
||
|
beq 9f // Nothing to reset!
|
||
|
|
||
|
/* Someone called kvm_call_hyp() against the hyp-stub... */
|
||
|
mov_q x0, HVC_STUB_ERR
|
||
|
eret
|
||
|
|
||
|
9: mov x0, xzr
|
||
|
eret
|
||
|
SYM_CODE_END(elx_sync)
|
||
|
|
||
|
SYM_CODE_START_LOCAL(__finalise_el2)
|
||
|
check_override id_aa64pfr0 ID_AA64PFR0_EL1_SVE_SHIFT .Linit_sve .Lskip_sve
|
||
|
|
||
|
.Linit_sve: /* SVE register access */
|
||
|
mrs x0, cptr_el2 // Disable SVE traps
|
||
|
bic x0, x0, #CPTR_EL2_TZ
|
||
|
msr cptr_el2, x0
|
||
|
isb
|
||
|
mov x1, #ZCR_ELx_LEN_MASK // SVE: Enable full vector
|
||
|
msr_s SYS_ZCR_EL2, x1 // length for EL1.
|
||
|
|
||
|
.Lskip_sve:
|
||
|
check_override id_aa64pfr1 ID_AA64PFR1_EL1_SME_SHIFT .Linit_sme .Lskip_sme
|
||
|
|
||
|
.Linit_sme: /* SME register access and priority mapping */
|
||
|
mrs x0, cptr_el2 // Disable SME traps
|
||
|
bic x0, x0, #CPTR_EL2_TSM
|
||
|
msr cptr_el2, x0
|
||
|
isb
|
||
|
|
||
|
mrs x1, sctlr_el2
|
||
|
orr x1, x1, #SCTLR_ELx_ENTP2 // Disable TPIDR2 traps
|
||
|
msr sctlr_el2, x1
|
||
|
isb
|
||
|
|
||
|
mov x0, #0 // SMCR controls
|
||
|
|
||
|
// Full FP in SM?
|
||
|
mrs_s x1, SYS_ID_AA64SMFR0_EL1
|
||
|
__check_override id_aa64smfr0 ID_AA64SMFR0_EL1_FA64_SHIFT 1 .Linit_sme_fa64 .Lskip_sme_fa64
|
||
|
|
||
|
.Linit_sme_fa64:
|
||
|
orr x0, x0, SMCR_ELx_FA64_MASK
|
||
|
.Lskip_sme_fa64:
|
||
|
|
||
|
orr x0, x0, #SMCR_ELx_LEN_MASK // Enable full SME vector
|
||
|
msr_s SYS_SMCR_EL2, x0 // length for EL1.
|
||
|
|
||
|
mrs_s x1, SYS_SMIDR_EL1 // Priority mapping supported?
|
||
|
ubfx x1, x1, #SMIDR_EL1_SMPS_SHIFT, #1
|
||
|
cbz x1, .Lskip_sme
|
||
|
|
||
|
msr_s SYS_SMPRIMAP_EL2, xzr // Make all priorities equal
|
||
|
|
||
|
mrs x1, id_aa64mmfr1_el1 // HCRX_EL2 present?
|
||
|
ubfx x1, x1, #ID_AA64MMFR1_EL1_HCX_SHIFT, #4
|
||
|
cbz x1, .Lskip_sme
|
||
|
|
||
|
mrs_s x1, SYS_HCRX_EL2
|
||
|
orr x1, x1, #HCRX_EL2_SMPME_MASK // Enable priority mapping
|
||
|
msr_s SYS_HCRX_EL2, x1
|
||
|
|
||
|
.Lskip_sme:
|
||
|
|
||
|
// nVHE? No way! Give me the real thing!
|
||
|
// Sanity check: MMU *must* be off
|
||
|
mrs x1, sctlr_el2
|
||
|
tbnz x1, #0, 1f
|
||
|
|
||
|
// Needs to be VHE capable, obviously
|
||
|
check_override id_aa64mmfr1 ID_AA64MMFR1_EL1_VH_SHIFT 2f 1f
|
||
|
|
||
|
1: mov_q x0, HVC_STUB_ERR
|
||
|
eret
|
||
|
2:
|
||
|
// Engage the VHE magic!
|
||
|
mov_q x0, HCR_HOST_VHE_FLAGS
|
||
|
msr hcr_el2, x0
|
||
|
isb
|
||
|
|
||
|
// Use the EL1 allocated stack, per-cpu offset
|
||
|
mrs x0, sp_el1
|
||
|
mov sp, x0
|
||
|
mrs x0, tpidr_el1
|
||
|
msr tpidr_el2, x0
|
||
|
|
||
|
// FP configuration, vectors
|
||
|
mrs_s x0, SYS_CPACR_EL12
|
||
|
msr cpacr_el1, x0
|
||
|
mrs_s x0, SYS_VBAR_EL12
|
||
|
msr vbar_el1, x0
|
||
|
|
||
|
// Use EL2 translations for SPE & TRBE and disable access from EL1
|
||
|
mrs x0, mdcr_el2
|
||
|
bic x0, x0, #(MDCR_EL2_E2PB_MASK << MDCR_EL2_E2PB_SHIFT)
|
||
|
bic x0, x0, #(MDCR_EL2_E2TB_MASK << MDCR_EL2_E2TB_SHIFT)
|
||
|
msr mdcr_el2, x0
|
||
|
|
||
|
// Transfer the MM state from EL1 to EL2
|
||
|
mrs_s x0, SYS_TCR_EL12
|
||
|
msr tcr_el1, x0
|
||
|
mrs_s x0, SYS_TTBR0_EL12
|
||
|
msr ttbr0_el1, x0
|
||
|
mrs_s x0, SYS_TTBR1_EL12
|
||
|
msr ttbr1_el1, x0
|
||
|
mrs_s x0, SYS_MAIR_EL12
|
||
|
msr mair_el1, x0
|
||
|
isb
|
||
|
|
||
|
// Hack the exception return to stay at EL2
|
||
|
mrs x0, spsr_el1
|
||
|
and x0, x0, #~PSR_MODE_MASK
|
||
|
mov x1, #PSR_MODE_EL2h
|
||
|
orr x0, x0, x1
|
||
|
msr spsr_el1, x0
|
||
|
|
||
|
b enter_vhe
|
||
|
SYM_CODE_END(__finalise_el2)
|
||
|
|
||
|
// At the point where we reach enter_vhe(), we run with
|
||
|
// the MMU off (which is enforced by __finalise_el2()).
|
||
|
// We thus need to be in the idmap, or everything will
|
||
|
// explode when enabling the MMU.
|
||
|
|
||
|
.pushsection .idmap.text, "ax"
|
||
|
|
||
|
SYM_CODE_START_LOCAL(enter_vhe)
|
||
|
// Invalidate TLBs before enabling the MMU
|
||
|
tlbi vmalle1
|
||
|
dsb nsh
|
||
|
isb
|
||
|
|
||
|
// Enable the EL2 S1 MMU, as set up from EL1
|
||
|
mrs_s x0, SYS_SCTLR_EL12
|
||
|
set_sctlr_el1 x0
|
||
|
|
||
|
// Disable the EL1 S1 MMU for a good measure
|
||
|
mov_q x0, INIT_SCTLR_EL1_MMU_OFF
|
||
|
msr_s SYS_SCTLR_EL12, x0
|
||
|
|
||
|
mov x0, xzr
|
||
|
|
||
|
eret
|
||
|
SYM_CODE_END(enter_vhe)
|
||
|
|
||
|
.popsection
|
||
|
|
||
|
.macro invalid_vector label
|
||
|
SYM_CODE_START_LOCAL(\label)
|
||
|
b \label
|
||
|
SYM_CODE_END(\label)
|
||
|
.endm
|
||
|
|
||
|
invalid_vector el2_sync_invalid
|
||
|
invalid_vector el2_irq_invalid
|
||
|
invalid_vector el2_fiq_invalid
|
||
|
invalid_vector el2_error_invalid
|
||
|
invalid_vector el1_sync_invalid
|
||
|
invalid_vector el1_irq_invalid
|
||
|
invalid_vector el1_fiq_invalid
|
||
|
invalid_vector el1_error_invalid
|
||
|
|
||
|
.popsection
|
||
|
|
||
|
/*
|
||
|
* __hyp_set_vectors: Call this after boot to set the initial hypervisor
|
||
|
* vectors as part of hypervisor installation. On an SMP system, this should
|
||
|
* be called on each CPU.
|
||
|
*
|
||
|
* x0 must be the physical address of the new vector table, and must be
|
||
|
* 2KB aligned.
|
||
|
*
|
||
|
* Before calling this, you must check that the stub hypervisor is installed
|
||
|
* everywhere, by waiting for any secondary CPUs to be brought up and then
|
||
|
* checking that is_hyp_mode_available() is true.
|
||
|
*
|
||
|
* If not, there is a pre-existing hypervisor, some CPUs failed to boot, or
|
||
|
* something else went wrong... in such cases, trying to install a new
|
||
|
* hypervisor is unlikely to work as desired.
|
||
|
*
|
||
|
* When you call into your shiny new hypervisor, sp_el2 will contain junk,
|
||
|
* so you will need to set that to something sensible at the new hypervisor's
|
||
|
* initialisation entry point.
|
||
|
*/
|
||
|
|
||
|
SYM_FUNC_START(__hyp_set_vectors)
|
||
|
mov x1, x0
|
||
|
mov x0, #HVC_SET_VECTORS
|
||
|
hvc #0
|
||
|
ret
|
||
|
SYM_FUNC_END(__hyp_set_vectors)
|
||
|
|
||
|
SYM_FUNC_START(__hyp_reset_vectors)
|
||
|
mov x0, #HVC_RESET_VECTORS
|
||
|
hvc #0
|
||
|
ret
|
||
|
SYM_FUNC_END(__hyp_reset_vectors)
|
||
|
|
||
|
/*
|
||
|
* Entry point to finalise EL2 and switch to VHE if deemed capable
|
||
|
*
|
||
|
* w0: boot mode, as returned by init_kernel_el()
|
||
|
*/
|
||
|
SYM_FUNC_START(finalise_el2)
|
||
|
// Need to have booted at EL2
|
||
|
cmp w0, #BOOT_CPU_MODE_EL2
|
||
|
b.ne 1f
|
||
|
|
||
|
// and still be at EL1
|
||
|
mrs x0, CurrentEL
|
||
|
cmp x0, #CurrentEL_EL1
|
||
|
b.ne 1f
|
||
|
|
||
|
mov x0, #HVC_FINALISE_EL2
|
||
|
hvc #0
|
||
|
1:
|
||
|
ret
|
||
|
SYM_FUNC_END(finalise_el2)
|