OvmfPkg/SmmRelocationLib: Add library instance for OVMF
There are below 2 differences between AMD & OVMF according existing implementation: 1.The mode of the CPU check is different between the AMD & OVMF. OVMF: CpuSaveState->x86.SMMRevId & 0Xffff AMD: LMAValue = (UINT32)AsmReadMsr64 (EFER_ADDRESS) & LMA 2.Existing SmBase configuration is different between the AMD & OVMF. OVMF: if ((CpuSaveState->x86.SMMRevId & 0xFFFF) == 0) { CpuSaveState->x86.SMBASE = mSmBaseForAllCpus[CpuIndex]; } else { CpuSaveState->x64.SMBASE = mSmBaseForAllCpus[CpuIndex]; } AMD: AmdCpuState->x64.SMBASE = mSmBaseForAllCpus[CpuIndex]; This patch provides the SmmRelocationLib library instance for OVMF to handle the logic difference, and it won't change the existing implementation code logic. Cc: Ray Ni <ray.ni@intel.com> Cc: Zeng Star <star.zeng@intel.com> Cc: Ard Biesheuvel <ardb+tianocore@kernel.org> Cc: Jiewen Yao <jiewen.yao@intel.com> Cc: Gerd Hoffmann <kraxel@redhat.com> Cc: Rahul Kumar <rahul1.kumar@intel.com> Signed-off-by: Jiaxin Wu <jiaxin.wu@intel.com> Tested-by: Gerd Hoffmann <kraxel@redhat.com> Acked-by: Gerd Hoffmann <kraxel@redhat.com> Acked-by: Jiewen Yao <Jiewen.yao@intel.com>
This commit is contained in:
66
OvmfPkg/Library/SmmRelocationLib/X64/Semaphore.c
Normal file
66
OvmfPkg/Library/SmmRelocationLib/X64/Semaphore.c
Normal file
@@ -0,0 +1,66 @@
|
||||
/** @file
|
||||
Semaphore mechanism to indicate to the BSP that an AP has exited SMM
|
||||
after SMBASE relocation.
|
||||
|
||||
Copyright (c) 2024, Intel Corporation. All rights reserved.<BR>
|
||||
SPDX-License-Identifier: BSD-2-Clause-Patent
|
||||
|
||||
**/
|
||||
|
||||
#include "InternalSmmRelocationLib.h"
|
||||
|
||||
X86_ASSEMBLY_PATCH_LABEL gPatchSmmRelocationOriginalAddressPtr32;
|
||||
X86_ASSEMBLY_PATCH_LABEL gPatchRebasedFlagAddr32;
|
||||
|
||||
UINTN mSmmRelocationOriginalAddress;
|
||||
volatile BOOLEAN *mRebasedFlag;
|
||||
|
||||
/**
|
||||
AP Semaphore operation in 32-bit mode while BSP runs in 64-bit mode.
|
||||
**/
|
||||
VOID
|
||||
SmmRelocationSemaphoreComplete32 (
|
||||
VOID
|
||||
);
|
||||
|
||||
/**
|
||||
Hook return address of SMM Save State so that semaphore code
|
||||
can be executed immediately after AP exits SMM to indicate to
|
||||
the BSP that an AP has exited SMM after SMBASE relocation.
|
||||
|
||||
@param[in] RebasedFlag A pointer to a flag that is set to TRUE
|
||||
immediately after AP exits SMM.
|
||||
|
||||
**/
|
||||
VOID
|
||||
SemaphoreHook (
|
||||
IN volatile BOOLEAN *RebasedFlag
|
||||
)
|
||||
{
|
||||
SMRAM_SAVE_STATE_MAP *CpuState;
|
||||
UINTN TempValue;
|
||||
|
||||
mRebasedFlag = RebasedFlag;
|
||||
PatchInstructionX86 (
|
||||
gPatchRebasedFlagAddr32,
|
||||
(UINT32)(UINTN)mRebasedFlag,
|
||||
4
|
||||
);
|
||||
|
||||
CpuState = (SMRAM_SAVE_STATE_MAP *)(UINTN)(SMM_DEFAULT_SMBASE + SMRAM_SAVE_STATE_MAP_OFFSET);
|
||||
mSmmRelocationOriginalAddress = HookReturnFromSmm (
|
||||
CpuState,
|
||||
(UINT64)(UINTN)&SmmRelocationSemaphoreComplete32,
|
||||
(UINT64)(UINTN)&SmmRelocationSemaphoreComplete
|
||||
);
|
||||
|
||||
//
|
||||
// Use temp value to fix ICC compiler warning
|
||||
//
|
||||
TempValue = (UINTN)&mSmmRelocationOriginalAddress;
|
||||
PatchInstructionX86 (
|
||||
gPatchSmmRelocationOriginalAddressPtr32,
|
||||
(UINT32)TempValue,
|
||||
4
|
||||
);
|
||||
}
|
201
OvmfPkg/Library/SmmRelocationLib/X64/SmmInit.nasm
Normal file
201
OvmfPkg/Library/SmmRelocationLib/X64/SmmInit.nasm
Normal file
@@ -0,0 +1,201 @@
|
||||
;------------------------------------------------------------------------------ ;
|
||||
; Copyright (c) 2024, Intel Corporation. All rights reserved.<BR>
|
||||
; SPDX-License-Identifier: BSD-2-Clause-Patent
|
||||
;
|
||||
; Module Name:
|
||||
;
|
||||
; SmmInit.nasm
|
||||
;
|
||||
; Abstract:
|
||||
;
|
||||
; Functions for relocating SMBASE's for all processors
|
||||
;
|
||||
;-------------------------------------------------------------------------------
|
||||
|
||||
%include "StuffRsbNasm.inc"
|
||||
|
||||
extern ASM_PFX(SmmInitHandler)
|
||||
extern ASM_PFX(mRebasedFlag)
|
||||
extern ASM_PFX(mSmmRelocationOriginalAddress)
|
||||
|
||||
global ASM_PFX(gPatchSmmInitCr3)
|
||||
global ASM_PFX(gPatchSmmInitCr4)
|
||||
global ASM_PFX(gPatchSmmInitCr0)
|
||||
global ASM_PFX(gPatchSmmInitStack)
|
||||
global ASM_PFX(gcSmmInitGdtr)
|
||||
global ASM_PFX(gcSmmInitSize)
|
||||
global ASM_PFX(gcSmmInitTemplate)
|
||||
global ASM_PFX(gPatchRebasedFlagAddr32)
|
||||
global ASM_PFX(gPatchSmmRelocationOriginalAddressPtr32)
|
||||
|
||||
%define LONG_MODE_CS 0x38
|
||||
|
||||
SECTION .data
|
||||
|
||||
NullSeg: DQ 0 ; reserved by architecture
|
||||
CodeSeg32:
|
||||
DW -1 ; LimitLow
|
||||
DW 0 ; BaseLow
|
||||
DB 0 ; BaseMid
|
||||
DB 0x9b
|
||||
DB 0xcf ; LimitHigh
|
||||
DB 0 ; BaseHigh
|
||||
ProtModeCodeSeg32:
|
||||
DW -1 ; LimitLow
|
||||
DW 0 ; BaseLow
|
||||
DB 0 ; BaseMid
|
||||
DB 0x9b
|
||||
DB 0xcf ; LimitHigh
|
||||
DB 0 ; BaseHigh
|
||||
ProtModeSsSeg32:
|
||||
DW -1 ; LimitLow
|
||||
DW 0 ; BaseLow
|
||||
DB 0 ; BaseMid
|
||||
DB 0x93
|
||||
DB 0xcf ; LimitHigh
|
||||
DB 0 ; BaseHigh
|
||||
DataSeg32:
|
||||
DW -1 ; LimitLow
|
||||
DW 0 ; BaseLow
|
||||
DB 0 ; BaseMid
|
||||
DB 0x93
|
||||
DB 0xcf ; LimitHigh
|
||||
DB 0 ; BaseHigh
|
||||
CodeSeg16:
|
||||
DW -1
|
||||
DW 0
|
||||
DB 0
|
||||
DB 0x9b
|
||||
DB 0x8f
|
||||
DB 0
|
||||
DataSeg16:
|
||||
DW -1
|
||||
DW 0
|
||||
DB 0
|
||||
DB 0x93
|
||||
DB 0x8f
|
||||
DB 0
|
||||
CodeSeg64:
|
||||
DW -1 ; LimitLow
|
||||
DW 0 ; BaseLow
|
||||
DB 0 ; BaseMid
|
||||
DB 0x9b
|
||||
DB 0xaf ; LimitHigh
|
||||
DB 0 ; BaseHigh
|
||||
GDT_SIZE equ $ - NullSeg
|
||||
|
||||
ASM_PFX(gcSmmInitGdtr):
|
||||
DW GDT_SIZE - 1
|
||||
DQ NullSeg
|
||||
|
||||
|
||||
DEFAULT REL
|
||||
SECTION .text
|
||||
|
||||
global ASM_PFX(SmmStartup)
|
||||
|
||||
BITS 16
|
||||
ASM_PFX(SmmStartup):
|
||||
mov eax, 0x80000001 ; read capability
|
||||
cpuid
|
||||
mov ebx, edx ; rdmsr will change edx. keep it in ebx.
|
||||
mov eax, strict dword 0 ; source operand will be patched
|
||||
ASM_PFX(gPatchSmmInitCr3):
|
||||
mov cr3, eax
|
||||
o32 lgdt [cs:ebp + (ASM_PFX(gcSmmInitGdtr) - ASM_PFX(SmmStartup))]
|
||||
mov eax, strict dword 0 ; source operand will be patched
|
||||
ASM_PFX(gPatchSmmInitCr4):
|
||||
or ah, 2 ; enable XMM registers access
|
||||
mov cr4, eax
|
||||
mov ecx, 0xc0000080 ; IA32_EFER MSR
|
||||
rdmsr
|
||||
or ah, BIT0 ; set LME bit
|
||||
test ebx, BIT20 ; check NXE capability
|
||||
jz .1
|
||||
or ah, BIT3 ; set NXE bit
|
||||
.1:
|
||||
wrmsr
|
||||
mov eax, strict dword 0 ; source operand will be patched
|
||||
ASM_PFX(gPatchSmmInitCr0):
|
||||
mov cr0, eax ; enable protected mode & paging
|
||||
jmp LONG_MODE_CS : dword 0 ; offset will be patched to @LongMode
|
||||
@PatchLongModeOffset:
|
||||
|
||||
BITS 64
|
||||
@LongMode: ; long-mode starts here
|
||||
mov rsp, strict qword 0 ; source operand will be patched
|
||||
ASM_PFX(gPatchSmmInitStack):
|
||||
and sp, 0xfff0 ; make sure RSP is 16-byte aligned
|
||||
;
|
||||
; According to X64 calling convention, XMM0~5 are volatile, we need to save
|
||||
; them before calling C-function.
|
||||
;
|
||||
sub rsp, 0x60
|
||||
movdqa [rsp], xmm0
|
||||
movdqa [rsp + 0x10], xmm1
|
||||
movdqa [rsp + 0x20], xmm2
|
||||
movdqa [rsp + 0x30], xmm3
|
||||
movdqa [rsp + 0x40], xmm4
|
||||
movdqa [rsp + 0x50], xmm5
|
||||
|
||||
add rsp, -0x20
|
||||
call ASM_PFX(SmmInitHandler)
|
||||
add rsp, 0x20
|
||||
|
||||
;
|
||||
; Restore XMM0~5 after calling C-function.
|
||||
;
|
||||
movdqa xmm0, [rsp]
|
||||
movdqa xmm1, [rsp + 0x10]
|
||||
movdqa xmm2, [rsp + 0x20]
|
||||
movdqa xmm3, [rsp + 0x30]
|
||||
movdqa xmm4, [rsp + 0x40]
|
||||
movdqa xmm5, [rsp + 0x50]
|
||||
|
||||
StuffRsb64
|
||||
rsm
|
||||
|
||||
BITS 16
|
||||
ASM_PFX(gcSmmInitTemplate):
|
||||
mov ebp, [cs:@L1 - ASM_PFX(gcSmmInitTemplate) + 0x8000]
|
||||
sub ebp, 0x30000
|
||||
jmp ebp
|
||||
@L1:
|
||||
DQ 0; ASM_PFX(SmmStartup)
|
||||
|
||||
ASM_PFX(gcSmmInitSize): DW $ - ASM_PFX(gcSmmInitTemplate)
|
||||
|
||||
BITS 64
|
||||
global ASM_PFX(SmmRelocationSemaphoreComplete)
|
||||
ASM_PFX(SmmRelocationSemaphoreComplete):
|
||||
push rax
|
||||
mov rax, [ASM_PFX(mRebasedFlag)]
|
||||
mov byte [rax], 1
|
||||
pop rax
|
||||
jmp [ASM_PFX(mSmmRelocationOriginalAddress)]
|
||||
|
||||
;
|
||||
; Semaphore code running in 32-bit mode
|
||||
;
|
||||
BITS 32
|
||||
global ASM_PFX(SmmRelocationSemaphoreComplete32)
|
||||
ASM_PFX(SmmRelocationSemaphoreComplete32):
|
||||
push eax
|
||||
mov eax, strict dword 0 ; source operand will be patched
|
||||
ASM_PFX(gPatchRebasedFlagAddr32):
|
||||
mov byte [eax], 1
|
||||
pop eax
|
||||
jmp dword [dword 0] ; destination will be patched
|
||||
ASM_PFX(gPatchSmmRelocationOriginalAddressPtr32):
|
||||
|
||||
BITS 64
|
||||
global ASM_PFX(SmmInitFixupAddress)
|
||||
ASM_PFX(SmmInitFixupAddress):
|
||||
lea rax, [@LongMode]
|
||||
lea rcx, [@PatchLongModeOffset - 6]
|
||||
mov dword [rcx], eax
|
||||
|
||||
lea rax, [ASM_PFX(SmmStartup)]
|
||||
lea rcx, [@L1]
|
||||
mov qword [rcx], rax
|
||||
ret
|
Reference in New Issue
Block a user