__FUNCTION__ is a pre-standard extension that gcc and Visual C++ among others support, while __func__ was standardized in C99. Since it's more standard, replace __FUNCTION__ with __func__ throughout OvmfPkg. Signed-off-by: Rebecca Cran <rebecca@bsdio.com> Reviewed-by: Michael D Kinney <michael.d.kinney@intel.com> Reviewed-by: Ard Biesheuvel <ardb@kernel.org> Reviewed-by: Sunil V L <sunilvl@ventanamicro.com>
		
			
				
	
	
		
			955 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			955 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /** @file
 | |
|   x64-specifc functionality for Page Table Setup.
 | |
| 
 | |
| Copyright (c) 2006 - 2020, Intel Corporation. All rights reserved.<BR>
 | |
| SPDX-License-Identifier: BSD-2-Clause-Patent
 | |
| **/
 | |
| 
 | |
| #include <Uefi/UefiBaseType.h>
 | |
| #include <Uefi/UefiSpec.h>
 | |
| #include <Pi/PiBootMode.h>
 | |
| #include <Pi/PiHob.h>
 | |
| #include <Library/DebugLib.h>
 | |
| #include <Library/BaseLib.h>
 | |
| #include <Library/HobLib.h>
 | |
| #include <Library/BaseMemoryLib.h>
 | |
| #include <Library/MemoryAllocationLib.h>
 | |
| #include <Library/PcdLib.h>
 | |
| #include <Guid/MemoryTypeInformation.h>
 | |
| #include <Guid/MemoryAllocationHob.h>
 | |
| #include <Register/Intel/Cpuid.h>
 | |
| #include <Library/PlatformInitLib.h>
 | |
| #include "PageTables.h"
 | |
| 
 | |
| UINTN  mLevelShift[5] = {
 | |
|   0,
 | |
|   PAGING_L1_ADDRESS_SHIFT,
 | |
|   PAGING_L2_ADDRESS_SHIFT,
 | |
|   PAGING_L3_ADDRESS_SHIFT,
 | |
|   PAGING_L4_ADDRESS_SHIFT
 | |
| };
 | |
| 
 | |
| UINT64  mLevelMask[5] = {
 | |
|   0,
 | |
|   PAGING_4K_ADDRESS_MASK_64,
 | |
|   PAGING_2M_ADDRESS_MASK_64,
 | |
|   PAGING_1G_ADDRESS_MASK_64,
 | |
|   PAGING_1G_ADDRESS_MASK_64
 | |
| };
 | |
| 
 | |
| UINT64  mLevelSize[5] = {
 | |
|   0,
 | |
|   SIZE_4KB,
 | |
|   SIZE_2MB,
 | |
|   SIZE_1GB,
 | |
|   SIZE_512GB
 | |
| };
 | |
| 
 | |
| BOOLEAN
 | |
| IsSetNxForStack (
 | |
|   VOID
 | |
|   )
 | |
| {
 | |
|   EFI_HOB_GUID_TYPE      *GuidHob;
 | |
|   EFI_HOB_PLATFORM_INFO  *PlatformInfo;
 | |
| 
 | |
|   GuidHob = GetFirstGuidHob (&gUefiOvmfPkgPlatformInfoGuid);
 | |
|   if (GuidHob == NULL) {
 | |
|     ASSERT (FALSE);
 | |
|     return FALSE;
 | |
|   }
 | |
| 
 | |
|   PlatformInfo = (EFI_HOB_PLATFORM_INFO *)GET_GUID_HOB_DATA (GuidHob);
 | |
| 
 | |
|   return PlatformInfo->PcdSetNxForStack;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Clear legacy memory located at the first 4K-page, if available.
 | |
| 
 | |
|   This function traverses the whole HOB list to check if memory from 0 to 4095
 | |
|   exists and has not been allocated, and then clear it if so.
 | |
| 
 | |
|   @param HobStart                  The start of HobList passed to DxeCore.
 | |
| 
 | |
| **/
 | |
| VOID
 | |
| ClearFirst4KPage (
 | |
|   IN  VOID  *HobStart
 | |
|   )
 | |
| {
 | |
|   EFI_PEI_HOB_POINTERS  RscHob;
 | |
|   EFI_PEI_HOB_POINTERS  MemHob;
 | |
|   BOOLEAN               DoClear;
 | |
| 
 | |
|   RscHob.Raw = HobStart;
 | |
|   MemHob.Raw = HobStart;
 | |
|   DoClear    = FALSE;
 | |
| 
 | |
|   //
 | |
|   // Check if page 0 exists and free
 | |
|   //
 | |
|   while ((RscHob.Raw = GetNextHob (
 | |
|                          EFI_HOB_TYPE_RESOURCE_DESCRIPTOR,
 | |
|                          RscHob.Raw
 | |
|                          )) != NULL)
 | |
|   {
 | |
|     if ((RscHob.ResourceDescriptor->ResourceType == EFI_RESOURCE_SYSTEM_MEMORY) &&
 | |
|         (RscHob.ResourceDescriptor->PhysicalStart == 0))
 | |
|     {
 | |
|       DoClear = TRUE;
 | |
|       //
 | |
|       // Make sure memory at 0-4095 has not been allocated.
 | |
|       //
 | |
|       while ((MemHob.Raw = GetNextHob (
 | |
|                              EFI_HOB_TYPE_MEMORY_ALLOCATION,
 | |
|                              MemHob.Raw
 | |
|                              )) != NULL)
 | |
|       {
 | |
|         if (MemHob.MemoryAllocation->AllocDescriptor.MemoryBaseAddress
 | |
|             < EFI_PAGE_SIZE)
 | |
|         {
 | |
|           DoClear = FALSE;
 | |
|           break;
 | |
|         }
 | |
| 
 | |
|         MemHob.Raw = GET_NEXT_HOB (MemHob);
 | |
|       }
 | |
| 
 | |
|       break;
 | |
|     }
 | |
| 
 | |
|     RscHob.Raw = GET_NEXT_HOB (RscHob);
 | |
|   }
 | |
| 
 | |
|   if (DoClear) {
 | |
|     DEBUG ((DEBUG_INFO, "Clearing first 4K-page!\r\n"));
 | |
|     SetMem (NULL, EFI_PAGE_SIZE, 0);
 | |
|   }
 | |
| 
 | |
|   return;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Return configure status of NULL pointer detection feature.
 | |
| 
 | |
|   @return TRUE   NULL pointer detection feature is enabled
 | |
|   @return FALSE  NULL pointer detection feature is disabled
 | |
| 
 | |
| **/
 | |
| BOOLEAN
 | |
| IsNullDetectionEnabled (
 | |
|   VOID
 | |
|   )
 | |
| {
 | |
|   return ((PcdGet8 (PcdNullPointerDetectionPropertyMask) & BIT0) != 0);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   The function will check if Execute Disable Bit is available.
 | |
| 
 | |
|   @retval TRUE      Execute Disable Bit is available.
 | |
|   @retval FALSE     Execute Disable Bit is not available.
 | |
| 
 | |
| **/
 | |
| BOOLEAN
 | |
| IsExecuteDisableBitAvailable (
 | |
|   VOID
 | |
|   )
 | |
| {
 | |
|   UINT32   RegEax;
 | |
|   UINT32   RegEdx;
 | |
|   BOOLEAN  Available;
 | |
| 
 | |
|   Available = FALSE;
 | |
|   AsmCpuid (0x80000000, &RegEax, NULL, NULL, NULL);
 | |
|   if (RegEax >= 0x80000001) {
 | |
|     AsmCpuid (0x80000001, NULL, NULL, NULL, &RegEdx);
 | |
|     if ((RegEdx & BIT20) != 0) {
 | |
|       //
 | |
|       // Bit 20: Execute Disable Bit available.
 | |
|       //
 | |
|       Available = TRUE;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   return Available;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Check if Execute Disable Bit (IA32_EFER.NXE) should be enabled or not.
 | |
| 
 | |
|   @retval TRUE    IA32_EFER.NXE should be enabled.
 | |
|   @retval FALSE   IA32_EFER.NXE should not be enabled.
 | |
| 
 | |
| **/
 | |
| BOOLEAN
 | |
| IsEnableNonExecNeeded (
 | |
|   VOID
 | |
|   )
 | |
| {
 | |
|   if (!IsExecuteDisableBitAvailable ()) {
 | |
|     return FALSE;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // XD flag (BIT63) in page table entry is only valid if IA32_EFER.NXE is set.
 | |
|   // Features controlled by Following PCDs need this feature to be enabled.
 | |
|   //
 | |
|   return (IsSetNxForStack () ||
 | |
|           FixedPcdGet64 (PcdDxeNxMemoryProtectionPolicy) != 0 ||
 | |
|           PcdGet32 (PcdImageProtectionPolicy) != 0);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Enable Execute Disable Bit.
 | |
| 
 | |
| **/
 | |
| VOID
 | |
| EnableExecuteDisableBit (
 | |
|   VOID
 | |
|   )
 | |
| {
 | |
|   UINT64  MsrRegisters;
 | |
| 
 | |
|   MsrRegisters  = AsmReadMsr64 (0xC0000080);
 | |
|   MsrRegisters |= BIT11;
 | |
|   AsmWriteMsr64 (0xC0000080, MsrRegisters);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   The function will check if page table entry should be splitted to smaller
 | |
|   granularity.
 | |
| 
 | |
|   @param Address      Physical memory address.
 | |
|   @param Size         Size of the given physical memory.
 | |
|   @param StackBase    Base address of stack.
 | |
|   @param StackSize    Size of stack.
 | |
| 
 | |
|   @retval TRUE      Page table should be split.
 | |
|   @retval FALSE     Page table should not be split.
 | |
| **/
 | |
| BOOLEAN
 | |
| ToSplitPageTable (
 | |
|   IN EFI_PHYSICAL_ADDRESS  Address,
 | |
|   IN UINTN                 Size,
 | |
|   IN EFI_PHYSICAL_ADDRESS  StackBase,
 | |
|   IN UINTN                 StackSize
 | |
|   )
 | |
| {
 | |
|   if (IsNullDetectionEnabled () && (Address == 0)) {
 | |
|     return TRUE;
 | |
|   }
 | |
| 
 | |
|   if (FixedPcdGetBool (PcdCpuStackGuard)) {
 | |
|     if ((StackBase >= Address) && (StackBase < (Address + Size))) {
 | |
|       return TRUE;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   if (IsSetNxForStack ()) {
 | |
|     if ((Address < StackBase + StackSize) && ((Address + Size) > StackBase)) {
 | |
|       return TRUE;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   return FALSE;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Initialize a buffer pool for page table use only.
 | |
| 
 | |
|   To reduce the potential split operation on page table, the pages reserved for
 | |
|   page table should be allocated in the times of PAGE_TABLE_POOL_UNIT_PAGES and
 | |
|   at the boundary of PAGE_TABLE_POOL_ALIGNMENT. So the page pool is always
 | |
|   initialized with number of pages greater than or equal to the given PoolPages.
 | |
| 
 | |
|   Once the pages in the pool are used up, this method should be called again to
 | |
|   reserve at least another PAGE_TABLE_POOL_UNIT_PAGES. But usually this won't
 | |
|   happen in practice.
 | |
| 
 | |
|   @param[in]      PoolPages      The least page number of the pool to be created.
 | |
|   @param[in, out] PageTablePool  Pointer of Pointer to the current available memory
 | |
|                                   used as page table.
 | |
| 
 | |
|   @retval TRUE    The pool is initialized successfully.
 | |
|   @retval FALSE   The memory is out of resource.
 | |
| **/
 | |
| BOOLEAN
 | |
| InitializePageTablePool (
 | |
|   IN UINTN                PoolPages,
 | |
|   IN OUT PAGE_TABLE_POOL  **PageTablePool
 | |
|   )
 | |
| {
 | |
|   VOID  *Buffer;
 | |
| 
 | |
|   DEBUG ((DEBUG_INFO, "InitializePageTablePool PoolPages=%d\n", PoolPages));
 | |
| 
 | |
|   //
 | |
|   // Always reserve at least PAGE_TABLE_POOL_UNIT_PAGES, including one page for
 | |
|   // header.
 | |
|   //
 | |
|   PoolPages += 1;   // Add one page for header.
 | |
|   PoolPages  = ((PoolPages - 1) / PAGE_TABLE_POOL_UNIT_PAGES + 1) *
 | |
|                PAGE_TABLE_POOL_UNIT_PAGES;
 | |
|   Buffer = AllocateAlignedPages (PoolPages, PAGE_TABLE_POOL_ALIGNMENT);
 | |
|   if (Buffer == NULL) {
 | |
|     DEBUG ((DEBUG_ERROR, "ERROR: Out of aligned pages\r\n"));
 | |
|     return FALSE;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Link all pools into a list for easier track later.
 | |
|   //
 | |
|   if (*PageTablePool == NULL) {
 | |
|     *(UINT64 *)(UINTN)PageTablePool = (UINT64)(UINTN)Buffer;
 | |
|     (*PageTablePool)->NextPool      = *PageTablePool;
 | |
|   } else {
 | |
|     ((PAGE_TABLE_POOL *)Buffer)->NextPool = (*PageTablePool)->NextPool;
 | |
|     (*PageTablePool)->NextPool            = Buffer;
 | |
|     *PageTablePool                        = Buffer;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Reserve one page for pool header.
 | |
|   //
 | |
|   (*PageTablePool)->FreePages = PoolPages - 1;
 | |
|   (*PageTablePool)->Offset    = EFI_PAGES_TO_SIZE (1);
 | |
| 
 | |
|   return TRUE;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   This API provides a way to allocate memory for page table.
 | |
| 
 | |
|   This API can be called more than once to allocate memory for page tables.
 | |
| 
 | |
|   Allocates the number of 4KB pages and returns a pointer to the allocated
 | |
|   buffer. The buffer returned is aligned on a 4KB boundary.
 | |
| 
 | |
|   If Pages is 0, then NULL is returned.
 | |
|   If there is not enough memory remaining to satisfy the request, then NULL is
 | |
|   returned.
 | |
| 
 | |
|   @param[in]      Pages                 The number of 4 KB pages to allocate.
 | |
|   @param[in, out] PageTablePool         Pointer of pointer to the current available
 | |
|                                         memory used as page table.
 | |
| 
 | |
|   @return A pointer to the allocated buffer or NULL if allocation fails.
 | |
| 
 | |
| **/
 | |
| VOID *
 | |
| AllocatePageTableMemory (
 | |
|   IN UINTN                Pages,
 | |
|   IN OUT PAGE_TABLE_POOL  **PageTablePool
 | |
|   )
 | |
| {
 | |
|   VOID  *Buffer;
 | |
| 
 | |
|   if (Pages == 0) {
 | |
|     return NULL;
 | |
|   }
 | |
| 
 | |
|   DEBUG ((DEBUG_INFO, "AllocatePageTableMemory. PageTablePool=%p, Pages=%d\n", *PageTablePool, Pages));
 | |
|   //
 | |
|   // Renew the pool if necessary.
 | |
|   //
 | |
|   if ((*PageTablePool == NULL) ||
 | |
|       (Pages > (*PageTablePool)->FreePages))
 | |
|   {
 | |
|     if (!InitializePageTablePool (Pages, PageTablePool)) {
 | |
|       return NULL;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   Buffer = (UINT8 *)(*PageTablePool) + (*PageTablePool)->Offset;
 | |
| 
 | |
|   (*PageTablePool)->Offset    += EFI_PAGES_TO_SIZE (Pages);
 | |
|   (*PageTablePool)->FreePages -= Pages;
 | |
| 
 | |
|   DEBUG ((
 | |
|     DEBUG_INFO,
 | |
|     "%a:%a: Buffer=0x%Lx Pages=%ld, PageTablePool=%p\n",
 | |
|     gEfiCallerBaseName,
 | |
|     __func__,
 | |
|     Buffer,
 | |
|     Pages,
 | |
|     *PageTablePool
 | |
|     ));
 | |
| 
 | |
|   return Buffer;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Split 2M page to 4K.
 | |
| 
 | |
|   @param[in]      PhysicalAddress       Start physical address the 2M page covered.
 | |
|   @param[in, out] PageEntry2M           Pointer to 2M page entry.
 | |
|   @param[in]      StackBase             Stack base address.
 | |
|   @param[in]      StackSize             Stack size.
 | |
|   @param[in, out] PageTablePool         Pointer to the current available memory used as
 | |
|                                         page table.
 | |
| 
 | |
| **/
 | |
| VOID
 | |
| Split2MPageTo4K (
 | |
|   IN EFI_PHYSICAL_ADDRESS  PhysicalAddress,
 | |
|   IN OUT UINT64            *PageEntry2M,
 | |
|   IN EFI_PHYSICAL_ADDRESS  StackBase,
 | |
|   IN UINTN                 StackSize,
 | |
|   IN OUT PAGE_TABLE_POOL   *PageTablePool
 | |
|   )
 | |
| {
 | |
|   EFI_PHYSICAL_ADDRESS  PhysicalAddress4K;
 | |
|   UINTN                 IndexOfPageTableEntries;
 | |
|   PAGE_TABLE_4K_ENTRY   *PageTableEntry;
 | |
| 
 | |
|   DEBUG ((DEBUG_INFO, "Split2MPageTo4K\n"));
 | |
| 
 | |
|   PageTableEntry = AllocatePageTableMemory (1, &PageTablePool);
 | |
| 
 | |
|   if (PageTableEntry == NULL) {
 | |
|     ASSERT (FALSE);
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Fill in 2M page entry.
 | |
|   //
 | |
|   *PageEntry2M = (UINT64)(UINTN)PageTableEntry | IA32_PG_P | IA32_PG_RW;
 | |
| 
 | |
|   PhysicalAddress4K = PhysicalAddress;
 | |
|   for (IndexOfPageTableEntries = 0; IndexOfPageTableEntries < 512; IndexOfPageTableEntries++, PageTableEntry++, PhysicalAddress4K += SIZE_4KB) {
 | |
|     //
 | |
|     // Fill in the Page Table entries
 | |
|     //
 | |
|     PageTableEntry->Uint64         = (UINT64)PhysicalAddress4K;
 | |
|     PageTableEntry->Bits.ReadWrite = 1;
 | |
| 
 | |
|     if ((IsNullDetectionEnabled () && (PhysicalAddress4K == 0)) ||
 | |
|         (FixedPcdGetBool (PcdCpuStackGuard) && (PhysicalAddress4K == StackBase)))
 | |
|     {
 | |
|       PageTableEntry->Bits.Present = 0;
 | |
|     } else {
 | |
|       PageTableEntry->Bits.Present = 1;
 | |
|     }
 | |
| 
 | |
|     if (  IsSetNxForStack ()
 | |
|        && (PhysicalAddress4K >= StackBase)
 | |
|        && (PhysicalAddress4K < StackBase + StackSize))
 | |
|     {
 | |
|       //
 | |
|       // Set Nx bit for stack.
 | |
|       //
 | |
|       PageTableEntry->Bits.Nx = 1;
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Split 1G page to 2M.
 | |
| 
 | |
|   @param[in]      PhysicalAddress       Start physical address the 1G page covered.
 | |
|   @param[in, out] PageEntry1G           Pointer to 1G page entry.
 | |
|   @param[in]      StackBase             Stack base address.
 | |
|   @param[in]      StackSize             Stack size.
 | |
|   @param[in, out] PageTablePool         Pointer to the current available memory used as
 | |
|                                         page table.
 | |
| 
 | |
| **/
 | |
| VOID
 | |
| Split1GPageTo2M (
 | |
|   IN EFI_PHYSICAL_ADDRESS  PhysicalAddress,
 | |
|   IN OUT UINT64            *PageEntry1G,
 | |
|   IN EFI_PHYSICAL_ADDRESS  StackBase,
 | |
|   IN UINTN                 StackSize,
 | |
|   IN OUT PAGE_TABLE_POOL   *PageTablePool
 | |
|   )
 | |
| {
 | |
|   EFI_PHYSICAL_ADDRESS  PhysicalAddress2M;
 | |
|   UINTN                 IndexOfPageDirectoryEntries;
 | |
|   PAGE_TABLE_ENTRY      *PageDirectoryEntry;
 | |
| 
 | |
|   DEBUG ((DEBUG_INFO, "Split1GPageTo2M\n"));
 | |
|   PageDirectoryEntry = AllocatePageTableMemory (1, &PageTablePool);
 | |
| 
 | |
|   if (PageDirectoryEntry == NULL) {
 | |
|     ASSERT (FALSE);
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Fill in 1G page entry.
 | |
|   //
 | |
|   *PageEntry1G = (UINT64)(UINTN)PageDirectoryEntry | IA32_PG_P | IA32_PG_RW;
 | |
| 
 | |
|   PhysicalAddress2M = PhysicalAddress;
 | |
|   for (IndexOfPageDirectoryEntries = 0; IndexOfPageDirectoryEntries < 512; IndexOfPageDirectoryEntries++, PageDirectoryEntry++, PhysicalAddress2M += SIZE_2MB) {
 | |
|     if (ToSplitPageTable (PhysicalAddress2M, SIZE_2MB, StackBase, StackSize)) {
 | |
|       //
 | |
|       // Need to split this 2M page that covers NULL or stack range.
 | |
|       //
 | |
|       Split2MPageTo4K (PhysicalAddress2M, (UINT64 *)PageDirectoryEntry, StackBase, StackSize, PageTablePool);
 | |
|     } else {
 | |
|       //
 | |
|       // Fill in the Page Directory entries
 | |
|       //
 | |
|       PageDirectoryEntry->Uint64         = (UINT64)PhysicalAddress2M;
 | |
|       PageDirectoryEntry->Bits.ReadWrite = 1;
 | |
|       PageDirectoryEntry->Bits.Present   = 1;
 | |
|       PageDirectoryEntry->Bits.MustBe1   = 1;
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Set one page of page table pool memory to be read-only.
 | |
| 
 | |
|   @param[in]      PageTableBase    Base address of page table (CR3).
 | |
|   @param[in]      Address          Start address of a page to be set as read-only.
 | |
|   @param[in]      Level4Paging     Level 4 paging flag.
 | |
|   @param[in, out] PageTablePool    Pointer to the current available memory used as
 | |
|                                    page table.
 | |
| 
 | |
| **/
 | |
| VOID
 | |
| SetPageTablePoolReadOnly (
 | |
|   IN  UINTN                 PageTableBase,
 | |
|   IN  EFI_PHYSICAL_ADDRESS  Address,
 | |
|   IN  BOOLEAN               Level4Paging,
 | |
|   IN OUT PAGE_TABLE_POOL    *PageTablePool
 | |
|   )
 | |
| {
 | |
|   UINTN                 Index;
 | |
|   UINTN                 EntryIndex;
 | |
|   EFI_PHYSICAL_ADDRESS  PhysicalAddress;
 | |
|   UINT64                *PageTable;
 | |
|   UINT64                *NewPageTable;
 | |
|   UINT64                PageAttr;
 | |
|   UINTN                 Level;
 | |
|   UINT64                PoolUnitSize;
 | |
| 
 | |
|   if (PageTableBase == 0) {
 | |
|     ASSERT (FALSE);
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Since the page table is always from page table pool, which is always
 | |
|   // located at the boundary of PcdPageTablePoolAlignment, we just need to
 | |
|   // set the whole pool unit to be read-only.
 | |
|   //
 | |
|   Address = Address & PAGE_TABLE_POOL_ALIGN_MASK;
 | |
| 
 | |
|   PageTable    = (UINT64 *)(UINTN)PageTableBase;
 | |
|   PoolUnitSize = PAGE_TABLE_POOL_UNIT_SIZE;
 | |
| 
 | |
|   for (Level = (Level4Paging) ? 4 : 3; Level > 0; --Level) {
 | |
|     Index  = ((UINTN)RShiftU64 (Address, mLevelShift[Level]));
 | |
|     Index &= PAGING_PAE_INDEX_MASK;
 | |
| 
 | |
|     PageAttr = PageTable[Index];
 | |
|     if ((PageAttr & IA32_PG_PS) == 0) {
 | |
|       //
 | |
|       // Go to next level of table.
 | |
|       //
 | |
|       PageTable = (UINT64 *)(UINTN)(PageAttr & PAGING_4K_ADDRESS_MASK_64);
 | |
|       continue;
 | |
|     }
 | |
| 
 | |
|     if (PoolUnitSize >= mLevelSize[Level]) {
 | |
|       //
 | |
|       // Clear R/W bit if current page granularity is not larger than pool unit
 | |
|       // size.
 | |
|       //
 | |
|       if ((PageAttr & IA32_PG_RW) != 0) {
 | |
|         while (PoolUnitSize > 0) {
 | |
|           //
 | |
|           // PAGE_TABLE_POOL_UNIT_SIZE and PAGE_TABLE_POOL_ALIGNMENT are fit in
 | |
|           // one page (2MB). Then we don't need to update attributes for pages
 | |
|           // crossing page directory. ASSERT below is for that purpose.
 | |
|           //
 | |
|           ASSERT (Index < EFI_PAGE_SIZE/sizeof (UINT64));
 | |
| 
 | |
|           PageTable[Index] &= ~(UINT64)IA32_PG_RW;
 | |
|           PoolUnitSize     -= mLevelSize[Level];
 | |
| 
 | |
|           ++Index;
 | |
|         }
 | |
|       }
 | |
| 
 | |
|       break;
 | |
|     } else {
 | |
|       //
 | |
|       // The smaller granularity of page must be needed.
 | |
|       //
 | |
|       ASSERT (Level > 1);
 | |
| 
 | |
|       DEBUG ((DEBUG_INFO, "SetPageTablePoolReadOnly\n"));
 | |
|       NewPageTable = AllocatePageTableMemory (1, &PageTablePool);
 | |
| 
 | |
|       if (NewPageTable == NULL) {
 | |
|         ASSERT (FALSE);
 | |
|         return;
 | |
|       }
 | |
| 
 | |
|       PhysicalAddress = PageAttr & mLevelMask[Level];
 | |
|       for (EntryIndex = 0;
 | |
|            EntryIndex < EFI_PAGE_SIZE/sizeof (UINT64);
 | |
|            ++EntryIndex)
 | |
|       {
 | |
|         NewPageTable[EntryIndex] = PhysicalAddress |
 | |
|                                    IA32_PG_P | IA32_PG_RW;
 | |
|         if (Level > 2) {
 | |
|           NewPageTable[EntryIndex] |= IA32_PG_PS;
 | |
|         }
 | |
| 
 | |
|         PhysicalAddress += mLevelSize[Level - 1];
 | |
|       }
 | |
| 
 | |
|       PageTable[Index] = (UINT64)(UINTN)NewPageTable |
 | |
|                          IA32_PG_P | IA32_PG_RW;
 | |
|       PageTable = NewPageTable;
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Prevent the memory pages used for page table from been overwritten.
 | |
| 
 | |
|   @param[in]      PageTableBase    Base address of page table (CR3).
 | |
|   @param[in]      Level4Paging     Level 4 paging flag.
 | |
|   @param[in, out] PageTablePool    Pointer to the current available memory used as
 | |
|                                    page table.
 | |
| 
 | |
| **/
 | |
| VOID
 | |
| EnablePageTableProtection (
 | |
|   IN  UINTN               PageTableBase,
 | |
|   IN  BOOLEAN             Level4Paging,
 | |
|   IN OUT PAGE_TABLE_POOL  *PageTablePool
 | |
|   )
 | |
| {
 | |
|   PAGE_TABLE_POOL       *HeadPool;
 | |
|   PAGE_TABLE_POOL       *Pool;
 | |
|   UINT64                PoolSize;
 | |
|   EFI_PHYSICAL_ADDRESS  Address;
 | |
| 
 | |
|   DEBUG ((DEBUG_INFO, "EnablePageTableProtection\n"));
 | |
| 
 | |
|   if (PageTablePool == NULL) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Disable write protection, because we need to mark page table to be write
 | |
|   // protected.
 | |
|   //
 | |
|   AsmWriteCr0 (AsmReadCr0 () & ~CR0_WP);
 | |
| 
 | |
|   //
 | |
|   // SetPageTablePoolReadOnly might update PageTablePool. It's safer to
 | |
|   // remember original one in advance.
 | |
|   //
 | |
|   HeadPool = PageTablePool;
 | |
|   Pool     = HeadPool;
 | |
|   do {
 | |
|     Address  = (EFI_PHYSICAL_ADDRESS)(UINTN)Pool;
 | |
|     PoolSize = Pool->Offset + EFI_PAGES_TO_SIZE (Pool->FreePages);
 | |
| 
 | |
|     //
 | |
|     // The size of one pool must be multiple of PAGE_TABLE_POOL_UNIT_SIZE, which
 | |
|     // is one of page size of the processor (2MB by default). Let's apply the
 | |
|     // protection to them one by one.
 | |
|     //
 | |
|     while (PoolSize > 0) {
 | |
|       SetPageTablePoolReadOnly (PageTableBase, Address, Level4Paging, PageTablePool);
 | |
|       Address  += PAGE_TABLE_POOL_UNIT_SIZE;
 | |
|       PoolSize -= PAGE_TABLE_POOL_UNIT_SIZE;
 | |
|     }
 | |
| 
 | |
|     Pool = Pool->NextPool;
 | |
|   } while (Pool != HeadPool);
 | |
| 
 | |
|   //
 | |
|   // Enable write protection, after page table attribute updated.
 | |
|   //
 | |
|   AsmWriteCr0 (AsmReadCr0 () | CR0_WP);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Allocates and fills in the Page Directory and Page Table Entries to
 | |
|   establish a 1:1 Virtual to Physical mapping.
 | |
| 
 | |
|   @param[in] StackBase  Stack base address.
 | |
|   @param[in] StackSize  Stack size.
 | |
| 
 | |
|   @return The address of 4 level page map.
 | |
| 
 | |
| **/
 | |
| UINTN
 | |
| CreateIdentityMappingPageTables (
 | |
|   IN EFI_PHYSICAL_ADDRESS  StackBase,
 | |
|   IN UINTN                 StackSize
 | |
|   )
 | |
| {
 | |
|   UINT32                          RegEax;
 | |
|   UINT32                          RegEdx;
 | |
|   UINT8                           PhysicalAddressBits;
 | |
|   EFI_PHYSICAL_ADDRESS            PageAddress;
 | |
|   UINTN                           IndexOfPml5Entries;
 | |
|   UINTN                           IndexOfPml4Entries;
 | |
|   UINTN                           IndexOfPdpEntries;
 | |
|   UINTN                           IndexOfPageDirectoryEntries;
 | |
|   UINT32                          NumberOfPml5EntriesNeeded;
 | |
|   UINT32                          NumberOfPml4EntriesNeeded;
 | |
|   UINT32                          NumberOfPdpEntriesNeeded;
 | |
|   PAGE_MAP_AND_DIRECTORY_POINTER  *PageMapLevel5Entry;
 | |
|   PAGE_MAP_AND_DIRECTORY_POINTER  *PageMapLevel4Entry;
 | |
|   PAGE_MAP_AND_DIRECTORY_POINTER  *PageMap;
 | |
|   PAGE_MAP_AND_DIRECTORY_POINTER  *PageDirectoryPointerEntry;
 | |
|   PAGE_TABLE_ENTRY                *PageDirectoryEntry;
 | |
|   UINTN                           TotalPagesNum;
 | |
|   UINTN                           BigPageAddress;
 | |
|   VOID                            *Hob;
 | |
|   BOOLEAN                         Page5LevelSupport;
 | |
|   BOOLEAN                         Page1GSupport;
 | |
|   PAGE_TABLE_1G_ENTRY             *PageDirectory1GEntry;
 | |
|   IA32_CR4                        Cr4;
 | |
|   PAGE_TABLE_POOL                 *PageTablePool;
 | |
| 
 | |
|   //
 | |
|   // Set PageMapLevel5Entry to suppress incorrect compiler/analyzer warnings
 | |
|   //
 | |
|   PageMapLevel5Entry = NULL;
 | |
| 
 | |
|   Page1GSupport = FALSE;
 | |
|   if (FixedPcdGetBool (PcdUse1GPageTable)) {
 | |
|     AsmCpuid (0x80000000, &RegEax, NULL, NULL, NULL);
 | |
|     if (RegEax >= 0x80000001) {
 | |
|       AsmCpuid (0x80000001, NULL, NULL, NULL, &RegEdx);
 | |
|       if ((RegEdx & BIT26) != 0) {
 | |
|         Page1GSupport = TRUE;
 | |
|       }
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Get physical address bits supported.
 | |
|   //
 | |
|   Hob = GetFirstHob (EFI_HOB_TYPE_CPU);
 | |
|   if (Hob == NULL) {
 | |
|     ASSERT (FALSE);
 | |
|     return 0;
 | |
|   }
 | |
| 
 | |
|   PhysicalAddressBits = ((EFI_HOB_CPU *)Hob)->SizeOfMemorySpace;
 | |
| 
 | |
|   //
 | |
|   // CPU will already have LA57 enabled so just check CR4
 | |
|   Cr4.UintN         = AsmReadCr4 ();
 | |
|   Page5LevelSupport = (Cr4.Bits.LA57 ? TRUE : FALSE);
 | |
| 
 | |
|   DEBUG ((
 | |
|     DEBUG_INFO,
 | |
|     "AddressBits=%u 5LevelPaging=%u 1GPage=%u \n",
 | |
|     PhysicalAddressBits,
 | |
|     Page5LevelSupport,
 | |
|     Page1GSupport
 | |
|     ));
 | |
| 
 | |
|   //
 | |
|   // Calculate the table entries needed.
 | |
|   //
 | |
|   NumberOfPml5EntriesNeeded = 1;
 | |
|   if (PhysicalAddressBits > 48) {
 | |
|     NumberOfPml5EntriesNeeded = (UINT32)LShiftU64 (1, PhysicalAddressBits - 48);
 | |
|     PhysicalAddressBits       = 48;
 | |
|   }
 | |
| 
 | |
|   NumberOfPml4EntriesNeeded = 1;
 | |
|   if (PhysicalAddressBits > 39) {
 | |
|     NumberOfPml4EntriesNeeded = (UINT32)LShiftU64 (1, PhysicalAddressBits - 39);
 | |
|     PhysicalAddressBits       = 39;
 | |
|   }
 | |
| 
 | |
|   NumberOfPdpEntriesNeeded = 1;
 | |
|   ASSERT (PhysicalAddressBits > 30);
 | |
|   NumberOfPdpEntriesNeeded = (UINT32)LShiftU64 (1, PhysicalAddressBits - 30);
 | |
| 
 | |
|   //
 | |
|   // Pre-allocate big pages to avoid later allocations.
 | |
|   //
 | |
|   if (!Page1GSupport) {
 | |
|     TotalPagesNum = ((NumberOfPdpEntriesNeeded + 1) * NumberOfPml4EntriesNeeded + 1) * NumberOfPml5EntriesNeeded + 1;
 | |
|   } else {
 | |
|     TotalPagesNum = (NumberOfPml4EntriesNeeded + 1) * NumberOfPml5EntriesNeeded + 1;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Substract the one page occupied by PML5 entries if 5-Level Paging is disabled.
 | |
|   //
 | |
|   if (!Page5LevelSupport) {
 | |
|     TotalPagesNum--;
 | |
|   }
 | |
| 
 | |
|   DEBUG ((
 | |
|     DEBUG_INFO,
 | |
|     "Pml5=%u Pml4=%u Pdp=%u TotalPage=%Lu\n",
 | |
|     NumberOfPml5EntriesNeeded,
 | |
|     NumberOfPml4EntriesNeeded,
 | |
|     NumberOfPdpEntriesNeeded,
 | |
|     (UINT64)TotalPagesNum
 | |
|     ));
 | |
| 
 | |
|   PageTablePool  = NULL;
 | |
|   BigPageAddress = (UINTN)AllocatePageTableMemory (TotalPagesNum, &PageTablePool);
 | |
|   if (BigPageAddress == 0) {
 | |
|     ASSERT (FALSE);
 | |
|     return 0;
 | |
|   }
 | |
| 
 | |
|   DEBUG ((DEBUG_INFO, "BigPageAddress = 0x%llx, PageTablePool=%p\n", BigPageAddress, PageTablePool));
 | |
| 
 | |
|   //
 | |
|   // By architecture only one PageMapLevel4 exists - so lets allocate storage for it.
 | |
|   //
 | |
|   PageMap = (VOID *)BigPageAddress;
 | |
|   if (Page5LevelSupport) {
 | |
|     //
 | |
|     // By architecture only one PageMapLevel5 exists - so lets allocate storage for it.
 | |
|     //
 | |
|     PageMapLevel5Entry = PageMap;
 | |
|     BigPageAddress    += SIZE_4KB;
 | |
|   }
 | |
| 
 | |
|   PageAddress = 0;
 | |
| 
 | |
|   for ( IndexOfPml5Entries = 0
 | |
|         ; IndexOfPml5Entries < NumberOfPml5EntriesNeeded
 | |
|         ; IndexOfPml5Entries++)
 | |
|   {
 | |
|     //
 | |
|     // Each PML5 entry points to a page of PML4 entires.
 | |
|     // So lets allocate space for them and fill them in the IndexOfPml4Entries loop.
 | |
|     // When 5-Level Paging is disabled, below allocation happens only once.
 | |
|     //
 | |
|     PageMapLevel4Entry = (VOID *)BigPageAddress;
 | |
|     BigPageAddress    += SIZE_4KB;
 | |
| 
 | |
|     if (Page5LevelSupport) {
 | |
|       //
 | |
|       // Make a PML5 Entry
 | |
|       //
 | |
|       PageMapLevel5Entry->Uint64         = (UINT64)(UINTN)PageMapLevel4Entry;
 | |
|       PageMapLevel5Entry->Bits.ReadWrite = 1;
 | |
|       PageMapLevel5Entry->Bits.Present   = 1;
 | |
|       PageMapLevel5Entry++;
 | |
|     }
 | |
| 
 | |
|     for ( IndexOfPml4Entries = 0
 | |
|           ; IndexOfPml4Entries < (NumberOfPml5EntriesNeeded == 1 ? NumberOfPml4EntriesNeeded : 512)
 | |
|           ; IndexOfPml4Entries++, PageMapLevel4Entry++)
 | |
|     {
 | |
|       //
 | |
|       // Each PML4 entry points to a page of Page Directory Pointer entires.
 | |
|       // So lets allocate space for them and fill them in the IndexOfPdpEntries loop.
 | |
|       //
 | |
|       PageDirectoryPointerEntry = (VOID *)BigPageAddress;
 | |
|       BigPageAddress           += SIZE_4KB;
 | |
| 
 | |
|       //
 | |
|       // Make a PML4 Entry
 | |
|       //
 | |
|       PageMapLevel4Entry->Uint64         = (UINT64)(UINTN)PageDirectoryPointerEntry;
 | |
|       PageMapLevel4Entry->Bits.ReadWrite = 1;
 | |
|       PageMapLevel4Entry->Bits.Present   = 1;
 | |
| 
 | |
|       if (Page1GSupport) {
 | |
|         PageDirectory1GEntry = (VOID *)PageDirectoryPointerEntry;
 | |
| 
 | |
|         for (IndexOfPageDirectoryEntries = 0; IndexOfPageDirectoryEntries < 512; IndexOfPageDirectoryEntries++, PageDirectory1GEntry++, PageAddress += SIZE_1GB) {
 | |
|           if (ToSplitPageTable (PageAddress, SIZE_1GB, StackBase, StackSize)) {
 | |
|             Split1GPageTo2M (
 | |
|               PageAddress,
 | |
|               (UINT64 *)PageDirectory1GEntry,
 | |
|               StackBase,
 | |
|               StackSize,
 | |
|               PageTablePool
 | |
|               );
 | |
|           } else {
 | |
|             //
 | |
|             // Fill in the Page Directory entries
 | |
|             //
 | |
|             PageDirectory1GEntry->Uint64         = (UINT64)PageAddress;
 | |
|             PageDirectory1GEntry->Bits.ReadWrite = 1;
 | |
|             PageDirectory1GEntry->Bits.Present   = 1;
 | |
|             PageDirectory1GEntry->Bits.MustBe1   = 1;
 | |
|           }
 | |
|         }
 | |
|       } else {
 | |
|         for ( IndexOfPdpEntries = 0
 | |
|               ; IndexOfPdpEntries < (NumberOfPml4EntriesNeeded == 1 ? NumberOfPdpEntriesNeeded : 512)
 | |
|               ; IndexOfPdpEntries++, PageDirectoryPointerEntry++)
 | |
|         {
 | |
|           //
 | |
|           // Each Directory Pointer entries points to a page of Page Directory entires.
 | |
|           // So allocate space for them and fill them in the IndexOfPageDirectoryEntries loop.
 | |
|           //
 | |
|           PageDirectoryEntry = (VOID *)BigPageAddress;
 | |
|           BigPageAddress    += SIZE_4KB;
 | |
| 
 | |
|           //
 | |
|           // Fill in a Page Directory Pointer Entries
 | |
|           //
 | |
|           PageDirectoryPointerEntry->Uint64         = (UINT64)(UINTN)PageDirectoryEntry;
 | |
|           PageDirectoryPointerEntry->Bits.ReadWrite = 1;
 | |
|           PageDirectoryPointerEntry->Bits.Present   = 1;
 | |
| 
 | |
|           for (IndexOfPageDirectoryEntries = 0; IndexOfPageDirectoryEntries < 512; IndexOfPageDirectoryEntries++, PageDirectoryEntry++, PageAddress += SIZE_2MB) {
 | |
|             if (ToSplitPageTable (PageAddress, SIZE_2MB, StackBase, StackSize)) {
 | |
|               //
 | |
|               // Need to split this 2M page that covers NULL or stack range.
 | |
|               //
 | |
|               Split2MPageTo4K (PageAddress, (UINT64 *)PageDirectoryEntry, StackBase, StackSize, PageTablePool);
 | |
|             } else {
 | |
|               //
 | |
|               // Fill in the Page Directory entries
 | |
|               //
 | |
|               PageDirectoryEntry->Uint64         = (UINT64)PageAddress;
 | |
|               PageDirectoryEntry->Bits.ReadWrite = 1;
 | |
|               PageDirectoryEntry->Bits.Present   = 1;
 | |
|               PageDirectoryEntry->Bits.MustBe1   = 1;
 | |
|             }
 | |
|           }
 | |
|         }
 | |
| 
 | |
|         //
 | |
|         // Fill with null entry for unused PDPTE
 | |
|         //
 | |
|         ZeroMem (PageDirectoryPointerEntry, (512 - IndexOfPdpEntries) * sizeof (PAGE_MAP_AND_DIRECTORY_POINTER));
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     //
 | |
|     // For the PML4 entries we are not using fill in a null entry.
 | |
|     //
 | |
|     ZeroMem (PageMapLevel4Entry, (512 - IndexOfPml4Entries) * sizeof (PAGE_MAP_AND_DIRECTORY_POINTER));
 | |
|   }
 | |
| 
 | |
|   if (Page5LevelSupport) {
 | |
|     //
 | |
|     // For the PML5 entries we are not using fill in a null entry.
 | |
|     //
 | |
|     ZeroMem (PageMapLevel5Entry, (512 - IndexOfPml5Entries) * sizeof (PAGE_MAP_AND_DIRECTORY_POINTER));
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Protect the page table by marking the memory used for page table to be
 | |
|   // read-only.
 | |
|   //
 | |
|   EnablePageTableProtection ((UINTN)PageMap, TRUE, PageTablePool);
 | |
| 
 | |
|   return (UINTN)PageMap;
 | |
| }
 |