https://bugzilla.tianocore.org/show_bug.cgi?id=1373 Replace BSD 2-Clause License with BSD+Patent License. This change is based on the following emails: https://lists.01.org/pipermail/edk2-devel/2019-February/036260.html https://lists.01.org/pipermail/edk2-devel/2018-October/030385.html RFCs with detailed process for the license change: V3: https://lists.01.org/pipermail/edk2-devel/2019-March/038116.html V2: https://lists.01.org/pipermail/edk2-devel/2019-March/037669.html V1: https://lists.01.org/pipermail/edk2-devel/2019-March/037500.html Contributed-under: TianoCore Contribution Agreement 1.1 Signed-off-by: Michael D Kinney <michael.d.kinney@intel.com> Reviewed-by: Hao Wu <hao.a.wu@intel.com> Reviewed-by: Jian J Wang <jian.j.wang@intel.com>
		
			
				
	
	
		
			1745 lines
		
	
	
		
			46 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1745 lines
		
	
	
		
			46 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /** @file
 | |
|   UEFI Heap Guard functions.
 | |
| 
 | |
| Copyright (c) 2017-2018, Intel Corporation. All rights reserved.<BR>
 | |
| SPDX-License-Identifier: BSD-2-Clause-Patent
 | |
| 
 | |
| **/
 | |
| 
 | |
| #include "DxeMain.h"
 | |
| #include "Imem.h"
 | |
| #include "HeapGuard.h"
 | |
| 
 | |
| //
 | |
| // Global to avoid infinite reentrance of memory allocation when updating
 | |
| // page table attributes, which may need allocate pages for new PDE/PTE.
 | |
| //
 | |
| GLOBAL_REMOVE_IF_UNREFERENCED BOOLEAN mOnGuarding = FALSE;
 | |
| 
 | |
| //
 | |
| // Pointer to table tracking the Guarded memory with bitmap, in which  '1'
 | |
| // is used to indicate memory guarded. '0' might be free memory or Guard
 | |
| // page itself, depending on status of memory adjacent to it.
 | |
| //
 | |
| GLOBAL_REMOVE_IF_UNREFERENCED UINT64 mGuardedMemoryMap = 0;
 | |
| 
 | |
| //
 | |
| // Current depth level of map table pointed by mGuardedMemoryMap.
 | |
| // mMapLevel must be initialized at least by 1. It will be automatically
 | |
| // updated according to the address of memory just tracked.
 | |
| //
 | |
| GLOBAL_REMOVE_IF_UNREFERENCED UINTN mMapLevel = 1;
 | |
| 
 | |
| //
 | |
| // Shift and mask for each level of map table
 | |
| //
 | |
| GLOBAL_REMOVE_IF_UNREFERENCED UINTN mLevelShift[GUARDED_HEAP_MAP_TABLE_DEPTH]
 | |
|                                     = GUARDED_HEAP_MAP_TABLE_DEPTH_SHIFTS;
 | |
| GLOBAL_REMOVE_IF_UNREFERENCED UINTN mLevelMask[GUARDED_HEAP_MAP_TABLE_DEPTH]
 | |
|                                     = GUARDED_HEAP_MAP_TABLE_DEPTH_MASKS;
 | |
| 
 | |
| //
 | |
| // Used for promoting freed but not used pages.
 | |
| //
 | |
| GLOBAL_REMOVE_IF_UNREFERENCED EFI_PHYSICAL_ADDRESS mLastPromotedPage = BASE_4GB;
 | |
| 
 | |
| /**
 | |
|   Set corresponding bits in bitmap table to 1 according to the address.
 | |
| 
 | |
|   @param[in]  Address     Start address to set for.
 | |
|   @param[in]  BitNumber   Number of bits to set.
 | |
|   @param[in]  BitMap      Pointer to bitmap which covers the Address.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| STATIC
 | |
| VOID
 | |
| SetBits (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Address,
 | |
|   IN UINTN                   BitNumber,
 | |
|   IN UINT64                  *BitMap
 | |
|   )
 | |
| {
 | |
|   UINTN           Lsbs;
 | |
|   UINTN           Qwords;
 | |
|   UINTN           Msbs;
 | |
|   UINTN           StartBit;
 | |
|   UINTN           EndBit;
 | |
| 
 | |
|   StartBit  = (UINTN)GUARDED_HEAP_MAP_ENTRY_BIT_INDEX (Address);
 | |
|   EndBit    = (StartBit + BitNumber - 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
| 
 | |
|   if ((StartBit + BitNumber) >= GUARDED_HEAP_MAP_ENTRY_BITS) {
 | |
|     Msbs    = (GUARDED_HEAP_MAP_ENTRY_BITS - StartBit) %
 | |
|               GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
|     Lsbs    = (EndBit + 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
|     Qwords  = (BitNumber - Msbs) / GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
|   } else {
 | |
|     Msbs    = BitNumber;
 | |
|     Lsbs    = 0;
 | |
|     Qwords  = 0;
 | |
|   }
 | |
| 
 | |
|   if (Msbs > 0) {
 | |
|     *BitMap |= LShiftU64 (LShiftU64 (1, Msbs) - 1, StartBit);
 | |
|     BitMap  += 1;
 | |
|   }
 | |
| 
 | |
|   if (Qwords > 0) {
 | |
|     SetMem64 ((VOID *)BitMap, Qwords * GUARDED_HEAP_MAP_ENTRY_BYTES,
 | |
|               (UINT64)-1);
 | |
|     BitMap += Qwords;
 | |
|   }
 | |
| 
 | |
|   if (Lsbs > 0) {
 | |
|     *BitMap |= (LShiftU64 (1, Lsbs) - 1);
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Set corresponding bits in bitmap table to 0 according to the address.
 | |
| 
 | |
|   @param[in]  Address     Start address to set for.
 | |
|   @param[in]  BitNumber   Number of bits to set.
 | |
|   @param[in]  BitMap      Pointer to bitmap which covers the Address.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| STATIC
 | |
| VOID
 | |
| ClearBits (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Address,
 | |
|   IN UINTN                   BitNumber,
 | |
|   IN UINT64                  *BitMap
 | |
|   )
 | |
| {
 | |
|   UINTN           Lsbs;
 | |
|   UINTN           Qwords;
 | |
|   UINTN           Msbs;
 | |
|   UINTN           StartBit;
 | |
|   UINTN           EndBit;
 | |
| 
 | |
|   StartBit  = (UINTN)GUARDED_HEAP_MAP_ENTRY_BIT_INDEX (Address);
 | |
|   EndBit    = (StartBit + BitNumber - 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
| 
 | |
|   if ((StartBit + BitNumber) >= GUARDED_HEAP_MAP_ENTRY_BITS) {
 | |
|     Msbs    = (GUARDED_HEAP_MAP_ENTRY_BITS - StartBit) %
 | |
|               GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
|     Lsbs    = (EndBit + 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
|     Qwords  = (BitNumber - Msbs) / GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
|   } else {
 | |
|     Msbs    = BitNumber;
 | |
|     Lsbs    = 0;
 | |
|     Qwords  = 0;
 | |
|   }
 | |
| 
 | |
|   if (Msbs > 0) {
 | |
|     *BitMap &= ~LShiftU64 (LShiftU64 (1, Msbs) - 1, StartBit);
 | |
|     BitMap  += 1;
 | |
|   }
 | |
| 
 | |
|   if (Qwords > 0) {
 | |
|     SetMem64 ((VOID *)BitMap, Qwords * GUARDED_HEAP_MAP_ENTRY_BYTES, 0);
 | |
|     BitMap += Qwords;
 | |
|   }
 | |
| 
 | |
|   if (Lsbs > 0) {
 | |
|     *BitMap &= ~(LShiftU64 (1, Lsbs) - 1);
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Get corresponding bits in bitmap table according to the address.
 | |
| 
 | |
|   The value of bit 0 corresponds to the status of memory at given Address.
 | |
|   No more than 64 bits can be retrieved in one call.
 | |
| 
 | |
|   @param[in]  Address     Start address to retrieve bits for.
 | |
|   @param[in]  BitNumber   Number of bits to get.
 | |
|   @param[in]  BitMap      Pointer to bitmap which covers the Address.
 | |
| 
 | |
|   @return An integer containing the bits information.
 | |
| **/
 | |
| STATIC
 | |
| UINT64
 | |
| GetBits (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Address,
 | |
|   IN UINTN                   BitNumber,
 | |
|   IN UINT64                  *BitMap
 | |
|   )
 | |
| {
 | |
|   UINTN           StartBit;
 | |
|   UINTN           EndBit;
 | |
|   UINTN           Lsbs;
 | |
|   UINTN           Msbs;
 | |
|   UINT64          Result;
 | |
| 
 | |
|   ASSERT (BitNumber <= GUARDED_HEAP_MAP_ENTRY_BITS);
 | |
| 
 | |
|   StartBit  = (UINTN)GUARDED_HEAP_MAP_ENTRY_BIT_INDEX (Address);
 | |
|   EndBit    = (StartBit + BitNumber - 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
| 
 | |
|   if ((StartBit + BitNumber) > GUARDED_HEAP_MAP_ENTRY_BITS) {
 | |
|     Msbs = GUARDED_HEAP_MAP_ENTRY_BITS - StartBit;
 | |
|     Lsbs = (EndBit + 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | |
|   } else {
 | |
|     Msbs = BitNumber;
 | |
|     Lsbs = 0;
 | |
|   }
 | |
| 
 | |
|   if (StartBit == 0 && BitNumber == GUARDED_HEAP_MAP_ENTRY_BITS) {
 | |
|     Result = *BitMap;
 | |
|   } else {
 | |
|     Result    = RShiftU64((*BitMap), StartBit) & (LShiftU64(1, Msbs) - 1);
 | |
|     if (Lsbs > 0) {
 | |
|       BitMap  += 1;
 | |
|       Result  |= LShiftU64 ((*BitMap) & (LShiftU64 (1, Lsbs) - 1), Msbs);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   return Result;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Locate the pointer of bitmap from the guarded memory bitmap tables, which
 | |
|   covers the given Address.
 | |
| 
 | |
|   @param[in]  Address       Start address to search the bitmap for.
 | |
|   @param[in]  AllocMapUnit  Flag to indicate memory allocation for the table.
 | |
|   @param[out] BitMap        Pointer to bitmap which covers the Address.
 | |
| 
 | |
|   @return The bit number from given Address to the end of current map table.
 | |
| **/
 | |
| UINTN
 | |
| FindGuardedMemoryMap (
 | |
|   IN  EFI_PHYSICAL_ADDRESS    Address,
 | |
|   IN  BOOLEAN                 AllocMapUnit,
 | |
|   OUT UINT64                  **BitMap
 | |
|   )
 | |
| {
 | |
|   UINTN                   Level;
 | |
|   UINT64                  *GuardMap;
 | |
|   UINT64                  MapMemory;
 | |
|   UINTN                   Index;
 | |
|   UINTN                   Size;
 | |
|   UINTN                   BitsToUnitEnd;
 | |
|   EFI_STATUS              Status;
 | |
| 
 | |
|   //
 | |
|   // Adjust current map table depth according to the address to access
 | |
|   //
 | |
|   while (AllocMapUnit &&
 | |
|          mMapLevel < GUARDED_HEAP_MAP_TABLE_DEPTH &&
 | |
|          RShiftU64 (
 | |
|            Address,
 | |
|            mLevelShift[GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel - 1]
 | |
|            ) != 0) {
 | |
| 
 | |
|     if (mGuardedMemoryMap != 0) {
 | |
|       Size = (mLevelMask[GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel - 1] + 1)
 | |
|              * GUARDED_HEAP_MAP_ENTRY_BYTES;
 | |
|       Status = CoreInternalAllocatePages (
 | |
|                   AllocateAnyPages,
 | |
|                   EfiBootServicesData,
 | |
|                   EFI_SIZE_TO_PAGES (Size),
 | |
|                   &MapMemory,
 | |
|                   FALSE
 | |
|                   );
 | |
|       ASSERT_EFI_ERROR (Status);
 | |
|       ASSERT (MapMemory != 0);
 | |
| 
 | |
|       SetMem ((VOID *)(UINTN)MapMemory, Size, 0);
 | |
| 
 | |
|       *(UINT64 *)(UINTN)MapMemory = mGuardedMemoryMap;
 | |
|       mGuardedMemoryMap = MapMemory;
 | |
|     }
 | |
| 
 | |
|     mMapLevel++;
 | |
| 
 | |
|   }
 | |
| 
 | |
|   GuardMap = &mGuardedMemoryMap;
 | |
|   for (Level = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | |
|        Level < GUARDED_HEAP_MAP_TABLE_DEPTH;
 | |
|        ++Level) {
 | |
| 
 | |
|     if (*GuardMap == 0) {
 | |
|       if (!AllocMapUnit) {
 | |
|         GuardMap = NULL;
 | |
|         break;
 | |
|       }
 | |
| 
 | |
|       Size = (mLevelMask[Level] + 1) * GUARDED_HEAP_MAP_ENTRY_BYTES;
 | |
|       Status = CoreInternalAllocatePages (
 | |
|                   AllocateAnyPages,
 | |
|                   EfiBootServicesData,
 | |
|                   EFI_SIZE_TO_PAGES (Size),
 | |
|                   &MapMemory,
 | |
|                   FALSE
 | |
|                   );
 | |
|       ASSERT_EFI_ERROR (Status);
 | |
|       ASSERT (MapMemory != 0);
 | |
| 
 | |
|       SetMem ((VOID *)(UINTN)MapMemory, Size, 0);
 | |
|       *GuardMap = MapMemory;
 | |
|     }
 | |
| 
 | |
|     Index     = (UINTN)RShiftU64 (Address, mLevelShift[Level]);
 | |
|     Index     &= mLevelMask[Level];
 | |
|     GuardMap  = (UINT64 *)(UINTN)((*GuardMap) + Index * sizeof (UINT64));
 | |
| 
 | |
|   }
 | |
| 
 | |
|   BitsToUnitEnd = GUARDED_HEAP_MAP_BITS - GUARDED_HEAP_MAP_BIT_INDEX (Address);
 | |
|   *BitMap       = GuardMap;
 | |
| 
 | |
|   return BitsToUnitEnd;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Set corresponding bits in bitmap table to 1 according to given memory range.
 | |
| 
 | |
|   @param[in]  Address       Memory address to guard from.
 | |
|   @param[in]  NumberOfPages Number of pages to guard.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| EFIAPI
 | |
| SetGuardedMemoryBits (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Address,
 | |
|   IN UINTN                   NumberOfPages
 | |
|   )
 | |
| {
 | |
|   UINT64            *BitMap;
 | |
|   UINTN             Bits;
 | |
|   UINTN             BitsToUnitEnd;
 | |
| 
 | |
|   while (NumberOfPages > 0) {
 | |
|     BitsToUnitEnd = FindGuardedMemoryMap (Address, TRUE, &BitMap);
 | |
|     ASSERT (BitMap != NULL);
 | |
| 
 | |
|     if (NumberOfPages > BitsToUnitEnd) {
 | |
|       // Cross map unit
 | |
|       Bits = BitsToUnitEnd;
 | |
|     } else {
 | |
|       Bits  = NumberOfPages;
 | |
|     }
 | |
| 
 | |
|     SetBits (Address, Bits, BitMap);
 | |
| 
 | |
|     NumberOfPages -= Bits;
 | |
|     Address       += EFI_PAGES_TO_SIZE (Bits);
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Clear corresponding bits in bitmap table according to given memory range.
 | |
| 
 | |
|   @param[in]  Address       Memory address to unset from.
 | |
|   @param[in]  NumberOfPages Number of pages to unset guard.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| EFIAPI
 | |
| ClearGuardedMemoryBits (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Address,
 | |
|   IN UINTN                   NumberOfPages
 | |
|   )
 | |
| {
 | |
|   UINT64            *BitMap;
 | |
|   UINTN             Bits;
 | |
|   UINTN             BitsToUnitEnd;
 | |
| 
 | |
|   while (NumberOfPages > 0) {
 | |
|     BitsToUnitEnd = FindGuardedMemoryMap (Address, TRUE, &BitMap);
 | |
|     ASSERT (BitMap != NULL);
 | |
| 
 | |
|     if (NumberOfPages > BitsToUnitEnd) {
 | |
|       // Cross map unit
 | |
|       Bits = BitsToUnitEnd;
 | |
|     } else {
 | |
|       Bits  = NumberOfPages;
 | |
|     }
 | |
| 
 | |
|     ClearBits (Address, Bits, BitMap);
 | |
| 
 | |
|     NumberOfPages -= Bits;
 | |
|     Address       += EFI_PAGES_TO_SIZE (Bits);
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Retrieve corresponding bits in bitmap table according to given memory range.
 | |
| 
 | |
|   @param[in]  Address       Memory address to retrieve from.
 | |
|   @param[in]  NumberOfPages Number of pages to retrieve.
 | |
| 
 | |
|   @return An integer containing the guarded memory bitmap.
 | |
| **/
 | |
| UINT64
 | |
| GetGuardedMemoryBits (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Address,
 | |
|   IN UINTN                   NumberOfPages
 | |
|   )
 | |
| {
 | |
|   UINT64            *BitMap;
 | |
|   UINTN             Bits;
 | |
|   UINT64            Result;
 | |
|   UINTN             Shift;
 | |
|   UINTN             BitsToUnitEnd;
 | |
| 
 | |
|   ASSERT (NumberOfPages <= GUARDED_HEAP_MAP_ENTRY_BITS);
 | |
| 
 | |
|   Result = 0;
 | |
|   Shift  = 0;
 | |
|   while (NumberOfPages > 0) {
 | |
|     BitsToUnitEnd = FindGuardedMemoryMap (Address, FALSE, &BitMap);
 | |
| 
 | |
|     if (NumberOfPages > BitsToUnitEnd) {
 | |
|       // Cross map unit
 | |
|       Bits  = BitsToUnitEnd;
 | |
|     } else {
 | |
|       Bits  = NumberOfPages;
 | |
|     }
 | |
| 
 | |
|     if (BitMap != NULL) {
 | |
|       Result |= LShiftU64 (GetBits (Address, Bits, BitMap), Shift);
 | |
|     }
 | |
| 
 | |
|     Shift         += Bits;
 | |
|     NumberOfPages -= Bits;
 | |
|     Address       += EFI_PAGES_TO_SIZE (Bits);
 | |
|   }
 | |
| 
 | |
|   return Result;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Get bit value in bitmap table for the given address.
 | |
| 
 | |
|   @param[in]  Address     The address to retrieve for.
 | |
| 
 | |
|   @return 1 or 0.
 | |
| **/
 | |
| UINTN
 | |
| EFIAPI
 | |
| GetGuardMapBit (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Address
 | |
|   )
 | |
| {
 | |
|   UINT64        *GuardMap;
 | |
| 
 | |
|   FindGuardedMemoryMap (Address, FALSE, &GuardMap);
 | |
|   if (GuardMap != NULL) {
 | |
|     if (RShiftU64 (*GuardMap,
 | |
|                    GUARDED_HEAP_MAP_ENTRY_BIT_INDEX (Address)) & 1) {
 | |
|       return 1;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   return 0;
 | |
| }
 | |
| 
 | |
| 
 | |
| /**
 | |
|   Check to see if the page at the given address is a Guard page or not.
 | |
| 
 | |
|   @param[in]  Address     The address to check for.
 | |
| 
 | |
|   @return TRUE  The page at Address is a Guard page.
 | |
|   @return FALSE The page at Address is not a Guard page.
 | |
| **/
 | |
| BOOLEAN
 | |
| EFIAPI
 | |
| IsGuardPage (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Address
 | |
|   )
 | |
| {
 | |
|   UINT64        BitMap;
 | |
| 
 | |
|   //
 | |
|   // There must be at least one guarded page before and/or after given
 | |
|   // address if it's a Guard page. The bitmap pattern should be one of
 | |
|   // 001, 100 and 101
 | |
|   //
 | |
|   BitMap = GetGuardedMemoryBits (Address - EFI_PAGE_SIZE, 3);
 | |
|   return ((BitMap == BIT0) || (BitMap == BIT2) || (BitMap == (BIT2 | BIT0)));
 | |
| }
 | |
| 
 | |
| 
 | |
| /**
 | |
|   Check to see if the page at the given address is guarded or not.
 | |
| 
 | |
|   @param[in]  Address     The address to check for.
 | |
| 
 | |
|   @return TRUE  The page at Address is guarded.
 | |
|   @return FALSE The page at Address is not guarded.
 | |
| **/
 | |
| BOOLEAN
 | |
| EFIAPI
 | |
| IsMemoryGuarded (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Address
 | |
|   )
 | |
| {
 | |
|   return (GetGuardMapBit (Address) == 1);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Set the page at the given address to be a Guard page.
 | |
| 
 | |
|   This is done by changing the page table attribute to be NOT PRSENT.
 | |
| 
 | |
|   @param[in]  BaseAddress     Page address to Guard at
 | |
| 
 | |
|   @return VOID
 | |
| **/
 | |
| VOID
 | |
| EFIAPI
 | |
| SetGuardPage (
 | |
|   IN  EFI_PHYSICAL_ADDRESS      BaseAddress
 | |
|   )
 | |
| {
 | |
|   EFI_STATUS      Status;
 | |
| 
 | |
|   if (gCpu == NULL) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Set flag to make sure allocating memory without GUARD for page table
 | |
|   // operation; otherwise infinite loops could be caused.
 | |
|   //
 | |
|   mOnGuarding = TRUE;
 | |
|   //
 | |
|   // Note: This might overwrite other attributes needed by other features,
 | |
|   // such as NX memory protection.
 | |
|   //
 | |
|   Status = gCpu->SetMemoryAttributes (gCpu, BaseAddress, EFI_PAGE_SIZE, EFI_MEMORY_RP);
 | |
|   ASSERT_EFI_ERROR (Status);
 | |
|   mOnGuarding = FALSE;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Unset the Guard page at the given address to the normal memory.
 | |
| 
 | |
|   This is done by changing the page table attribute to be PRSENT.
 | |
| 
 | |
|   @param[in]  BaseAddress     Page address to Guard at.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| EFIAPI
 | |
| UnsetGuardPage (
 | |
|   IN  EFI_PHYSICAL_ADDRESS      BaseAddress
 | |
|   )
 | |
| {
 | |
|   UINT64          Attributes;
 | |
|   EFI_STATUS      Status;
 | |
| 
 | |
|   if (gCpu == NULL) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Once the Guard page is unset, it will be freed back to memory pool. NX
 | |
|   // memory protection must be restored for this page if NX is enabled for free
 | |
|   // memory.
 | |
|   //
 | |
|   Attributes = 0;
 | |
|   if ((PcdGet64 (PcdDxeNxMemoryProtectionPolicy) & (1 << EfiConventionalMemory)) != 0) {
 | |
|     Attributes |= EFI_MEMORY_XP;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Set flag to make sure allocating memory without GUARD for page table
 | |
|   // operation; otherwise infinite loops could be caused.
 | |
|   //
 | |
|   mOnGuarding = TRUE;
 | |
|   //
 | |
|   // Note: This might overwrite other attributes needed by other features,
 | |
|   // such as memory protection (NX). Please make sure they are not enabled
 | |
|   // at the same time.
 | |
|   //
 | |
|   Status = gCpu->SetMemoryAttributes (gCpu, BaseAddress, EFI_PAGE_SIZE, Attributes);
 | |
|   ASSERT_EFI_ERROR (Status);
 | |
|   mOnGuarding = FALSE;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Check to see if the memory at the given address should be guarded or not.
 | |
| 
 | |
|   @param[in]  MemoryType      Memory type to check.
 | |
|   @param[in]  AllocateType    Allocation type to check.
 | |
|   @param[in]  PageOrPool      Indicate a page allocation or pool allocation.
 | |
| 
 | |
| 
 | |
|   @return TRUE  The given type of memory should be guarded.
 | |
|   @return FALSE The given type of memory should not be guarded.
 | |
| **/
 | |
| BOOLEAN
 | |
| IsMemoryTypeToGuard (
 | |
|   IN EFI_MEMORY_TYPE        MemoryType,
 | |
|   IN EFI_ALLOCATE_TYPE      AllocateType,
 | |
|   IN UINT8                  PageOrPool
 | |
|   )
 | |
| {
 | |
|   UINT64 TestBit;
 | |
|   UINT64 ConfigBit;
 | |
| 
 | |
|   if (AllocateType == AllocateAddress) {
 | |
|     return FALSE;
 | |
|   }
 | |
| 
 | |
|   if ((PcdGet8 (PcdHeapGuardPropertyMask) & PageOrPool) == 0) {
 | |
|     return FALSE;
 | |
|   }
 | |
| 
 | |
|   if (PageOrPool == GUARD_HEAP_TYPE_POOL) {
 | |
|     ConfigBit = PcdGet64 (PcdHeapGuardPoolType);
 | |
|   } else if (PageOrPool == GUARD_HEAP_TYPE_PAGE) {
 | |
|     ConfigBit = PcdGet64 (PcdHeapGuardPageType);
 | |
|   } else {
 | |
|     ConfigBit = (UINT64)-1;
 | |
|   }
 | |
| 
 | |
|   if ((UINT32)MemoryType >= MEMORY_TYPE_OS_RESERVED_MIN) {
 | |
|     TestBit = BIT63;
 | |
|   } else if ((UINT32) MemoryType >= MEMORY_TYPE_OEM_RESERVED_MIN) {
 | |
|     TestBit = BIT62;
 | |
|   } else if (MemoryType < EfiMaxMemoryType) {
 | |
|     TestBit = LShiftU64 (1, MemoryType);
 | |
|   } else if (MemoryType == EfiMaxMemoryType) {
 | |
|     TestBit = (UINT64)-1;
 | |
|   } else {
 | |
|     TestBit = 0;
 | |
|   }
 | |
| 
 | |
|   return ((ConfigBit & TestBit) != 0);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Check to see if the pool at the given address should be guarded or not.
 | |
| 
 | |
|   @param[in]  MemoryType      Pool type to check.
 | |
| 
 | |
| 
 | |
|   @return TRUE  The given type of pool should be guarded.
 | |
|   @return FALSE The given type of pool should not be guarded.
 | |
| **/
 | |
| BOOLEAN
 | |
| IsPoolTypeToGuard (
 | |
|   IN EFI_MEMORY_TYPE        MemoryType
 | |
|   )
 | |
| {
 | |
|   return IsMemoryTypeToGuard (MemoryType, AllocateAnyPages,
 | |
|                               GUARD_HEAP_TYPE_POOL);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Check to see if the page at the given address should be guarded or not.
 | |
| 
 | |
|   @param[in]  MemoryType      Page type to check.
 | |
|   @param[in]  AllocateType    Allocation type to check.
 | |
| 
 | |
|   @return TRUE  The given type of page should be guarded.
 | |
|   @return FALSE The given type of page should not be guarded.
 | |
| **/
 | |
| BOOLEAN
 | |
| IsPageTypeToGuard (
 | |
|   IN EFI_MEMORY_TYPE        MemoryType,
 | |
|   IN EFI_ALLOCATE_TYPE      AllocateType
 | |
|   )
 | |
| {
 | |
|   return IsMemoryTypeToGuard (MemoryType, AllocateType, GUARD_HEAP_TYPE_PAGE);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Check to see if the heap guard is enabled for page and/or pool allocation.
 | |
| 
 | |
|   @param[in]  GuardType   Specify the sub-type(s) of Heap Guard.
 | |
| 
 | |
|   @return TRUE/FALSE.
 | |
| **/
 | |
| BOOLEAN
 | |
| IsHeapGuardEnabled (
 | |
|   UINT8           GuardType
 | |
|   )
 | |
| {
 | |
|   return IsMemoryTypeToGuard (EfiMaxMemoryType, AllocateAnyPages, GuardType);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Set head Guard and tail Guard for the given memory range.
 | |
| 
 | |
|   @param[in]  Memory          Base address of memory to set guard for.
 | |
|   @param[in]  NumberOfPages   Memory size in pages.
 | |
| 
 | |
|   @return VOID
 | |
| **/
 | |
| VOID
 | |
| SetGuardForMemory (
 | |
|   IN EFI_PHYSICAL_ADDRESS   Memory,
 | |
|   IN UINTN                  NumberOfPages
 | |
|   )
 | |
| {
 | |
|   EFI_PHYSICAL_ADDRESS    GuardPage;
 | |
| 
 | |
|   //
 | |
|   // Set tail Guard
 | |
|   //
 | |
|   GuardPage = Memory + EFI_PAGES_TO_SIZE (NumberOfPages);
 | |
|   if (!IsGuardPage (GuardPage)) {
 | |
|     SetGuardPage (GuardPage);
 | |
|   }
 | |
| 
 | |
|   // Set head Guard
 | |
|   GuardPage = Memory - EFI_PAGES_TO_SIZE (1);
 | |
|   if (!IsGuardPage (GuardPage)) {
 | |
|     SetGuardPage (GuardPage);
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Mark the memory range as Guarded
 | |
|   //
 | |
|   SetGuardedMemoryBits (Memory, NumberOfPages);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Unset head Guard and tail Guard for the given memory range.
 | |
| 
 | |
|   @param[in]  Memory          Base address of memory to unset guard for.
 | |
|   @param[in]  NumberOfPages   Memory size in pages.
 | |
| 
 | |
|   @return VOID
 | |
| **/
 | |
| VOID
 | |
| UnsetGuardForMemory (
 | |
|   IN EFI_PHYSICAL_ADDRESS   Memory,
 | |
|   IN UINTN                  NumberOfPages
 | |
|   )
 | |
| {
 | |
|   EFI_PHYSICAL_ADDRESS  GuardPage;
 | |
|   UINT64                GuardBitmap;
 | |
| 
 | |
|   if (NumberOfPages == 0) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Head Guard must be one page before, if any.
 | |
|   //
 | |
|   //          MSB-> 1     0 <-LSB
 | |
|   //          -------------------
 | |
|   //  Head Guard -> 0     1 -> Don't free Head Guard  (shared Guard)
 | |
|   //  Head Guard -> 0     0 -> Free Head Guard either (not shared Guard)
 | |
|   //                1     X -> Don't free first page  (need a new Guard)
 | |
|   //                           (it'll be turned into a Guard page later)
 | |
|   //          -------------------
 | |
|   //      Start -> -1    -2
 | |
|   //
 | |
|   GuardPage = Memory - EFI_PAGES_TO_SIZE (1);
 | |
|   GuardBitmap = GetGuardedMemoryBits (Memory - EFI_PAGES_TO_SIZE (2), 2);
 | |
|   if ((GuardBitmap & BIT1) == 0) {
 | |
|     //
 | |
|     // Head Guard exists.
 | |
|     //
 | |
|     if ((GuardBitmap & BIT0) == 0) {
 | |
|       //
 | |
|       // If the head Guard is not a tail Guard of adjacent memory block,
 | |
|       // unset it.
 | |
|       //
 | |
|       UnsetGuardPage (GuardPage);
 | |
|     }
 | |
|   } else {
 | |
|     //
 | |
|     // Pages before memory to free are still in Guard. It's a partial free
 | |
|     // case. Turn first page of memory block to free into a new Guard.
 | |
|     //
 | |
|     SetGuardPage (Memory);
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Tail Guard must be the page after this memory block to free, if any.
 | |
|   //
 | |
|   //   MSB-> 1     0 <-LSB
 | |
|   //  --------------------
 | |
|   //         1     0 <- Tail Guard -> Don't free Tail Guard  (shared Guard)
 | |
|   //         0     0 <- Tail Guard -> Free Tail Guard either (not shared Guard)
 | |
|   //         X     1               -> Don't free last page   (need a new Guard)
 | |
|   //                                 (it'll be turned into a Guard page later)
 | |
|   //  --------------------
 | |
|   //        +1    +0 <- End
 | |
|   //
 | |
|   GuardPage = Memory + EFI_PAGES_TO_SIZE (NumberOfPages);
 | |
|   GuardBitmap = GetGuardedMemoryBits (GuardPage, 2);
 | |
|   if ((GuardBitmap & BIT0) == 0) {
 | |
|     //
 | |
|     // Tail Guard exists.
 | |
|     //
 | |
|     if ((GuardBitmap & BIT1) == 0) {
 | |
|       //
 | |
|       // If the tail Guard is not a head Guard of adjacent memory block,
 | |
|       // free it; otherwise, keep it.
 | |
|       //
 | |
|       UnsetGuardPage (GuardPage);
 | |
|     }
 | |
|   } else {
 | |
|     //
 | |
|     // Pages after memory to free are still in Guard. It's a partial free
 | |
|     // case. We need to keep one page to be a head Guard.
 | |
|     //
 | |
|     SetGuardPage (GuardPage - EFI_PAGES_TO_SIZE (1));
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // No matter what, we just clear the mark of the Guarded memory.
 | |
|   //
 | |
|   ClearGuardedMemoryBits(Memory, NumberOfPages);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Adjust address of free memory according to existing and/or required Guard.
 | |
| 
 | |
|   This function will check if there're existing Guard pages of adjacent
 | |
|   memory blocks, and try to use it as the Guard page of the memory to be
 | |
|   allocated.
 | |
| 
 | |
|   @param[in]  Start           Start address of free memory block.
 | |
|   @param[in]  Size            Size of free memory block.
 | |
|   @param[in]  SizeRequested   Size of memory to allocate.
 | |
| 
 | |
|   @return The end address of memory block found.
 | |
|   @return 0 if no enough space for the required size of memory and its Guard.
 | |
| **/
 | |
| UINT64
 | |
| AdjustMemoryS (
 | |
|   IN UINT64                  Start,
 | |
|   IN UINT64                  Size,
 | |
|   IN UINT64                  SizeRequested
 | |
|   )
 | |
| {
 | |
|   UINT64  Target;
 | |
| 
 | |
|   //
 | |
|   // UEFI spec requires that allocated pool must be 8-byte aligned. If it's
 | |
|   // indicated to put the pool near the Tail Guard, we need extra bytes to
 | |
|   // make sure alignment of the returned pool address.
 | |
|   //
 | |
|   if ((PcdGet8 (PcdHeapGuardPropertyMask) & BIT7) == 0) {
 | |
|     SizeRequested = ALIGN_VALUE(SizeRequested, 8);
 | |
|   }
 | |
| 
 | |
|   Target = Start + Size - SizeRequested;
 | |
|   ASSERT (Target >= Start);
 | |
|   if (Target == 0) {
 | |
|     return 0;
 | |
|   }
 | |
| 
 | |
|   if (!IsGuardPage (Start + Size)) {
 | |
|     // No Guard at tail to share. One more page is needed.
 | |
|     Target -= EFI_PAGES_TO_SIZE (1);
 | |
|   }
 | |
| 
 | |
|   // Out of range?
 | |
|   if (Target < Start) {
 | |
|     return 0;
 | |
|   }
 | |
| 
 | |
|   // At the edge?
 | |
|   if (Target == Start) {
 | |
|     if (!IsGuardPage (Target - EFI_PAGES_TO_SIZE (1))) {
 | |
|       // No enough space for a new head Guard if no Guard at head to share.
 | |
|       return 0;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   // OK, we have enough pages for memory and its Guards. Return the End of the
 | |
|   // free space.
 | |
|   return Target + SizeRequested - 1;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Adjust the start address and number of pages to free according to Guard.
 | |
| 
 | |
|   The purpose of this function is to keep the shared Guard page with adjacent
 | |
|   memory block if it's still in guard, or free it if no more sharing. Another
 | |
|   is to reserve pages as Guard pages in partial page free situation.
 | |
| 
 | |
|   @param[in,out]  Memory          Base address of memory to free.
 | |
|   @param[in,out]  NumberOfPages   Size of memory to free.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| AdjustMemoryF (
 | |
|   IN OUT EFI_PHYSICAL_ADDRESS    *Memory,
 | |
|   IN OUT UINTN                   *NumberOfPages
 | |
|   )
 | |
| {
 | |
|   EFI_PHYSICAL_ADDRESS  Start;
 | |
|   EFI_PHYSICAL_ADDRESS  MemoryToTest;
 | |
|   UINTN                 PagesToFree;
 | |
|   UINT64                GuardBitmap;
 | |
| 
 | |
|   if (Memory == NULL || NumberOfPages == NULL || *NumberOfPages == 0) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   Start = *Memory;
 | |
|   PagesToFree = *NumberOfPages;
 | |
| 
 | |
|   //
 | |
|   // Head Guard must be one page before, if any.
 | |
|   //
 | |
|   //          MSB-> 1     0 <-LSB
 | |
|   //          -------------------
 | |
|   //  Head Guard -> 0     1 -> Don't free Head Guard  (shared Guard)
 | |
|   //  Head Guard -> 0     0 -> Free Head Guard either (not shared Guard)
 | |
|   //                1     X -> Don't free first page  (need a new Guard)
 | |
|   //                           (it'll be turned into a Guard page later)
 | |
|   //          -------------------
 | |
|   //      Start -> -1    -2
 | |
|   //
 | |
|   MemoryToTest = Start - EFI_PAGES_TO_SIZE (2);
 | |
|   GuardBitmap = GetGuardedMemoryBits (MemoryToTest, 2);
 | |
|   if ((GuardBitmap & BIT1) == 0) {
 | |
|     //
 | |
|     // Head Guard exists.
 | |
|     //
 | |
|     if ((GuardBitmap & BIT0) == 0) {
 | |
|       //
 | |
|       // If the head Guard is not a tail Guard of adjacent memory block,
 | |
|       // free it; otherwise, keep it.
 | |
|       //
 | |
|       Start       -= EFI_PAGES_TO_SIZE (1);
 | |
|       PagesToFree += 1;
 | |
|     }
 | |
|   } else {
 | |
|     //
 | |
|     // No Head Guard, and pages before memory to free are still in Guard. It's a
 | |
|     // partial free case. We need to keep one page to be a tail Guard.
 | |
|     //
 | |
|     Start       += EFI_PAGES_TO_SIZE (1);
 | |
|     PagesToFree -= 1;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Tail Guard must be the page after this memory block to free, if any.
 | |
|   //
 | |
|   //   MSB-> 1     0 <-LSB
 | |
|   //  --------------------
 | |
|   //         1     0 <- Tail Guard -> Don't free Tail Guard  (shared Guard)
 | |
|   //         0     0 <- Tail Guard -> Free Tail Guard either (not shared Guard)
 | |
|   //         X     1               -> Don't free last page   (need a new Guard)
 | |
|   //                                 (it'll be turned into a Guard page later)
 | |
|   //  --------------------
 | |
|   //        +1    +0 <- End
 | |
|   //
 | |
|   MemoryToTest = Start + EFI_PAGES_TO_SIZE (PagesToFree);
 | |
|   GuardBitmap = GetGuardedMemoryBits (MemoryToTest, 2);
 | |
|   if ((GuardBitmap & BIT0) == 0) {
 | |
|     //
 | |
|     // Tail Guard exists.
 | |
|     //
 | |
|     if ((GuardBitmap & BIT1) == 0) {
 | |
|       //
 | |
|       // If the tail Guard is not a head Guard of adjacent memory block,
 | |
|       // free it; otherwise, keep it.
 | |
|       //
 | |
|       PagesToFree += 1;
 | |
|     }
 | |
|   } else if (PagesToFree > 0) {
 | |
|     //
 | |
|     // No Tail Guard, and pages after memory to free are still in Guard. It's a
 | |
|     // partial free case. We need to keep one page to be a head Guard.
 | |
|     //
 | |
|     PagesToFree -= 1;
 | |
|   }
 | |
| 
 | |
|   *Memory         = Start;
 | |
|   *NumberOfPages  = PagesToFree;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Adjust the base and number of pages to really allocate according to Guard.
 | |
| 
 | |
|   @param[in,out]  Memory          Base address of free memory.
 | |
|   @param[in,out]  NumberOfPages   Size of memory to allocate.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| AdjustMemoryA (
 | |
|   IN OUT EFI_PHYSICAL_ADDRESS    *Memory,
 | |
|   IN OUT UINTN                   *NumberOfPages
 | |
|   )
 | |
| {
 | |
|   //
 | |
|   // FindFreePages() has already taken the Guard into account. It's safe to
 | |
|   // adjust the start address and/or number of pages here, to make sure that
 | |
|   // the Guards are also "allocated".
 | |
|   //
 | |
|   if (!IsGuardPage (*Memory + EFI_PAGES_TO_SIZE (*NumberOfPages))) {
 | |
|     // No tail Guard, add one.
 | |
|     *NumberOfPages += 1;
 | |
|   }
 | |
| 
 | |
|   if (!IsGuardPage (*Memory - EFI_PAGE_SIZE)) {
 | |
|     // No head Guard, add one.
 | |
|     *Memory        -= EFI_PAGE_SIZE;
 | |
|     *NumberOfPages += 1;
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Adjust the pool head position to make sure the Guard page is adjavent to
 | |
|   pool tail or pool head.
 | |
| 
 | |
|   @param[in]  Memory    Base address of memory allocated.
 | |
|   @param[in]  NoPages   Number of pages actually allocated.
 | |
|   @param[in]  Size      Size of memory requested.
 | |
|                         (plus pool head/tail overhead)
 | |
| 
 | |
|   @return Address of pool head.
 | |
| **/
 | |
| VOID *
 | |
| AdjustPoolHeadA (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Memory,
 | |
|   IN UINTN                   NoPages,
 | |
|   IN UINTN                   Size
 | |
|   )
 | |
| {
 | |
|   if (Memory == 0 || (PcdGet8 (PcdHeapGuardPropertyMask) & BIT7) != 0) {
 | |
|     //
 | |
|     // Pool head is put near the head Guard
 | |
|     //
 | |
|     return (VOID *)(UINTN)Memory;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Pool head is put near the tail Guard
 | |
|   //
 | |
|   Size = ALIGN_VALUE (Size, 8);
 | |
|   return (VOID *)(UINTN)(Memory + EFI_PAGES_TO_SIZE (NoPages) - Size);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Get the page base address according to pool head address.
 | |
| 
 | |
|   @param[in]  Memory    Head address of pool to free.
 | |
| 
 | |
|   @return Address of pool head.
 | |
| **/
 | |
| VOID *
 | |
| AdjustPoolHeadF (
 | |
|   IN EFI_PHYSICAL_ADDRESS    Memory
 | |
|   )
 | |
| {
 | |
|   if (Memory == 0 || (PcdGet8 (PcdHeapGuardPropertyMask) & BIT7) != 0) {
 | |
|     //
 | |
|     // Pool head is put near the head Guard
 | |
|     //
 | |
|     return (VOID *)(UINTN)Memory;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Pool head is put near the tail Guard
 | |
|   //
 | |
|   return (VOID *)(UINTN)(Memory & ~EFI_PAGE_MASK);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Allocate or free guarded memory.
 | |
| 
 | |
|   @param[in]  Start           Start address of memory to allocate or free.
 | |
|   @param[in]  NumberOfPages   Memory size in pages.
 | |
|   @param[in]  NewType         Memory type to convert to.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| EFI_STATUS
 | |
| CoreConvertPagesWithGuard (
 | |
|   IN UINT64           Start,
 | |
|   IN UINTN            NumberOfPages,
 | |
|   IN EFI_MEMORY_TYPE  NewType
 | |
|   )
 | |
| {
 | |
|   UINT64  OldStart;
 | |
|   UINTN   OldPages;
 | |
| 
 | |
|   if (NewType == EfiConventionalMemory) {
 | |
|     OldStart = Start;
 | |
|     OldPages = NumberOfPages;
 | |
| 
 | |
|     AdjustMemoryF (&Start, &NumberOfPages);
 | |
|     //
 | |
|     // It's safe to unset Guard page inside memory lock because there should
 | |
|     // be no memory allocation occurred in updating memory page attribute at
 | |
|     // this point. And unsetting Guard page before free will prevent Guard
 | |
|     // page just freed back to pool from being allocated right away before
 | |
|     // marking it usable (from non-present to present).
 | |
|     //
 | |
|     UnsetGuardForMemory (OldStart, OldPages);
 | |
|     if (NumberOfPages == 0) {
 | |
|       return EFI_SUCCESS;
 | |
|     }
 | |
|   } else {
 | |
|     AdjustMemoryA (&Start, &NumberOfPages);
 | |
|   }
 | |
| 
 | |
|   return CoreConvertPages (Start, NumberOfPages, NewType);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Set all Guard pages which cannot be set before CPU Arch Protocol installed.
 | |
| **/
 | |
| VOID
 | |
| SetAllGuardPages (
 | |
|   VOID
 | |
|   )
 | |
| {
 | |
|   UINTN     Entries[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINTN     Shifts[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINTN     Indices[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINT64    Tables[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINT64    Addresses[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINT64    TableEntry;
 | |
|   UINT64    Address;
 | |
|   UINT64    GuardPage;
 | |
|   INTN      Level;
 | |
|   UINTN     Index;
 | |
|   BOOLEAN   OnGuarding;
 | |
| 
 | |
|   if (mGuardedMemoryMap == 0 ||
 | |
|       mMapLevel == 0 ||
 | |
|       mMapLevel > GUARDED_HEAP_MAP_TABLE_DEPTH) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   CopyMem (Entries, mLevelMask, sizeof (Entries));
 | |
|   CopyMem (Shifts, mLevelShift, sizeof (Shifts));
 | |
| 
 | |
|   SetMem (Tables, sizeof(Tables), 0);
 | |
|   SetMem (Addresses, sizeof(Addresses), 0);
 | |
|   SetMem (Indices, sizeof(Indices), 0);
 | |
| 
 | |
|   Level         = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | |
|   Tables[Level] = mGuardedMemoryMap;
 | |
|   Address       = 0;
 | |
|   OnGuarding    = FALSE;
 | |
| 
 | |
|   DEBUG_CODE (
 | |
|     DumpGuardedMemoryBitmap ();
 | |
|   );
 | |
| 
 | |
|   while (TRUE) {
 | |
|     if (Indices[Level] > Entries[Level]) {
 | |
|       Tables[Level] = 0;
 | |
|       Level        -= 1;
 | |
|     } else {
 | |
| 
 | |
|       TableEntry  = ((UINT64 *)(UINTN)(Tables[Level]))[Indices[Level]];
 | |
|       Address     = Addresses[Level];
 | |
| 
 | |
|       if (TableEntry == 0) {
 | |
| 
 | |
|         OnGuarding = FALSE;
 | |
| 
 | |
|       } else if (Level < GUARDED_HEAP_MAP_TABLE_DEPTH - 1) {
 | |
| 
 | |
|         Level            += 1;
 | |
|         Tables[Level]     = TableEntry;
 | |
|         Addresses[Level]  = Address;
 | |
|         Indices[Level]    = 0;
 | |
| 
 | |
|         continue;
 | |
| 
 | |
|       } else {
 | |
| 
 | |
|         Index = 0;
 | |
|         while (Index < GUARDED_HEAP_MAP_ENTRY_BITS) {
 | |
|           if ((TableEntry & 1) == 1) {
 | |
|             if (OnGuarding) {
 | |
|               GuardPage = 0;
 | |
|             } else {
 | |
|               GuardPage = Address - EFI_PAGE_SIZE;
 | |
|             }
 | |
|             OnGuarding = TRUE;
 | |
|           } else {
 | |
|             if (OnGuarding) {
 | |
|               GuardPage = Address;
 | |
|             } else {
 | |
|               GuardPage = 0;
 | |
|             }
 | |
|             OnGuarding = FALSE;
 | |
|           }
 | |
| 
 | |
|           if (GuardPage != 0) {
 | |
|             SetGuardPage (GuardPage);
 | |
|           }
 | |
| 
 | |
|           if (TableEntry == 0) {
 | |
|             break;
 | |
|           }
 | |
| 
 | |
|           TableEntry = RShiftU64 (TableEntry, 1);
 | |
|           Address   += EFI_PAGE_SIZE;
 | |
|           Index     += 1;
 | |
|         }
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     if (Level < (GUARDED_HEAP_MAP_TABLE_DEPTH - (INTN)mMapLevel)) {
 | |
|       break;
 | |
|     }
 | |
| 
 | |
|     Indices[Level] += 1;
 | |
|     Address = (Level == 0) ? 0 : Addresses[Level - 1];
 | |
|     Addresses[Level] = Address | LShiftU64(Indices[Level], Shifts[Level]);
 | |
| 
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Find the address of top-most guarded free page.
 | |
| 
 | |
|   @param[out]  Address    Start address of top-most guarded free page.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| GetLastGuardedFreePageAddress (
 | |
|   OUT EFI_PHYSICAL_ADDRESS      *Address
 | |
|   )
 | |
| {
 | |
|   EFI_PHYSICAL_ADDRESS    AddressGranularity;
 | |
|   EFI_PHYSICAL_ADDRESS    BaseAddress;
 | |
|   UINTN                   Level;
 | |
|   UINT64                  Map;
 | |
|   INTN                    Index;
 | |
| 
 | |
|   ASSERT (mMapLevel >= 1);
 | |
| 
 | |
|   BaseAddress = 0;
 | |
|   Map = mGuardedMemoryMap;
 | |
|   for (Level = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | |
|        Level < GUARDED_HEAP_MAP_TABLE_DEPTH;
 | |
|        ++Level) {
 | |
|     AddressGranularity = LShiftU64 (1, mLevelShift[Level]);
 | |
| 
 | |
|     //
 | |
|     // Find the non-NULL entry at largest index.
 | |
|     //
 | |
|     for (Index = (INTN)mLevelMask[Level]; Index >= 0 ; --Index) {
 | |
|       if (((UINT64 *)(UINTN)Map)[Index] != 0) {
 | |
|         BaseAddress += MultU64x32 (AddressGranularity, (UINT32)Index);
 | |
|         Map = ((UINT64 *)(UINTN)Map)[Index];
 | |
|         break;
 | |
|       }
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Find the non-zero MSB then get the page address.
 | |
|   //
 | |
|   while (Map != 0) {
 | |
|     Map = RShiftU64 (Map, 1);
 | |
|     BaseAddress += EFI_PAGES_TO_SIZE (1);
 | |
|   }
 | |
| 
 | |
|   *Address = BaseAddress;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Record freed pages.
 | |
| 
 | |
|   @param[in]  BaseAddress   Base address of just freed pages.
 | |
|   @param[in]  Pages         Number of freed pages.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| MarkFreedPages (
 | |
|   IN EFI_PHYSICAL_ADDRESS     BaseAddress,
 | |
|   IN UINTN                    Pages
 | |
|   )
 | |
| {
 | |
|   SetGuardedMemoryBits (BaseAddress, Pages);
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Record freed pages as well as mark them as not-present.
 | |
| 
 | |
|   @param[in]  BaseAddress   Base address of just freed pages.
 | |
|   @param[in]  Pages         Number of freed pages.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| EFIAPI
 | |
| GuardFreedPages (
 | |
|   IN  EFI_PHYSICAL_ADDRESS    BaseAddress,
 | |
|   IN  UINTN                   Pages
 | |
|   )
 | |
| {
 | |
|   EFI_STATUS      Status;
 | |
| 
 | |
|   //
 | |
|   // Legacy memory lower than 1MB might be accessed with no allocation. Leave
 | |
|   // them alone.
 | |
|   //
 | |
|   if (BaseAddress < BASE_1MB) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   MarkFreedPages (BaseAddress, Pages);
 | |
|   if (gCpu != NULL) {
 | |
|     //
 | |
|     // Set flag to make sure allocating memory without GUARD for page table
 | |
|     // operation; otherwise infinite loops could be caused.
 | |
|     //
 | |
|     mOnGuarding = TRUE;
 | |
|     //
 | |
|     // Note: This might overwrite other attributes needed by other features,
 | |
|     // such as NX memory protection.
 | |
|     //
 | |
|     Status = gCpu->SetMemoryAttributes (
 | |
|                      gCpu,
 | |
|                      BaseAddress,
 | |
|                      EFI_PAGES_TO_SIZE (Pages),
 | |
|                      EFI_MEMORY_RP
 | |
|                      );
 | |
|     //
 | |
|     // Normally we should ASSERT the returned Status. But there might be memory
 | |
|     // alloc/free involved in SetMemoryAttributes(), which might fail this
 | |
|     // calling. It's rare case so it's OK to let a few tiny holes be not-guarded.
 | |
|     //
 | |
|     if (EFI_ERROR (Status)) {
 | |
|       DEBUG ((DEBUG_WARN, "Failed to guard freed pages: %p (%lu)\n", BaseAddress, (UINT64)Pages));
 | |
|     }
 | |
|     mOnGuarding = FALSE;
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Record freed pages as well as mark them as not-present, if enabled.
 | |
| 
 | |
|   @param[in]  BaseAddress   Base address of just freed pages.
 | |
|   @param[in]  Pages         Number of freed pages.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| EFIAPI
 | |
| GuardFreedPagesChecked (
 | |
|   IN  EFI_PHYSICAL_ADDRESS    BaseAddress,
 | |
|   IN  UINTN                   Pages
 | |
|   )
 | |
| {
 | |
|   if (IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED)) {
 | |
|     GuardFreedPages (BaseAddress, Pages);
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Mark all pages freed before CPU Arch Protocol as not-present.
 | |
| 
 | |
| **/
 | |
| VOID
 | |
| GuardAllFreedPages (
 | |
|   VOID
 | |
|   )
 | |
| {
 | |
|   UINTN     Entries[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINTN     Shifts[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINTN     Indices[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINT64    Tables[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINT64    Addresses[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINT64    TableEntry;
 | |
|   UINT64    Address;
 | |
|   UINT64    GuardPage;
 | |
|   INTN      Level;
 | |
|   UINT64    BitIndex;
 | |
|   UINTN     GuardPageNumber;
 | |
| 
 | |
|   if (mGuardedMemoryMap == 0 ||
 | |
|       mMapLevel == 0 ||
 | |
|       mMapLevel > GUARDED_HEAP_MAP_TABLE_DEPTH) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   CopyMem (Entries, mLevelMask, sizeof (Entries));
 | |
|   CopyMem (Shifts, mLevelShift, sizeof (Shifts));
 | |
| 
 | |
|   SetMem (Tables, sizeof(Tables), 0);
 | |
|   SetMem (Addresses, sizeof(Addresses), 0);
 | |
|   SetMem (Indices, sizeof(Indices), 0);
 | |
| 
 | |
|   Level           = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | |
|   Tables[Level]   = mGuardedMemoryMap;
 | |
|   Address         = 0;
 | |
|   GuardPage       = (UINT64)-1;
 | |
|   GuardPageNumber = 0;
 | |
| 
 | |
|   while (TRUE) {
 | |
|     if (Indices[Level] > Entries[Level]) {
 | |
|       Tables[Level] = 0;
 | |
|       Level        -= 1;
 | |
|     } else {
 | |
|       TableEntry  = ((UINT64 *)(UINTN)(Tables[Level]))[Indices[Level]];
 | |
|       Address     = Addresses[Level];
 | |
| 
 | |
|       if (Level < GUARDED_HEAP_MAP_TABLE_DEPTH - 1) {
 | |
|         Level            += 1;
 | |
|         Tables[Level]     = TableEntry;
 | |
|         Addresses[Level]  = Address;
 | |
|         Indices[Level]    = 0;
 | |
| 
 | |
|         continue;
 | |
|       } else {
 | |
|         BitIndex = 1;
 | |
|         while (BitIndex != 0) {
 | |
|           if ((TableEntry & BitIndex) != 0) {
 | |
|             if (GuardPage == (UINT64)-1) {
 | |
|               GuardPage = Address;
 | |
|             }
 | |
|             ++GuardPageNumber;
 | |
|           } else if (GuardPageNumber > 0) {
 | |
|             GuardFreedPages (GuardPage, GuardPageNumber);
 | |
|             GuardPageNumber = 0;
 | |
|             GuardPage       = (UINT64)-1;
 | |
|           }
 | |
| 
 | |
|           if (TableEntry == 0) {
 | |
|             break;
 | |
|           }
 | |
| 
 | |
|           Address += EFI_PAGES_TO_SIZE (1);
 | |
|           BitIndex = LShiftU64 (BitIndex, 1);
 | |
|         }
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     if (Level < (GUARDED_HEAP_MAP_TABLE_DEPTH - (INTN)mMapLevel)) {
 | |
|       break;
 | |
|     }
 | |
| 
 | |
|     Indices[Level] += 1;
 | |
|     Address = (Level == 0) ? 0 : Addresses[Level - 1];
 | |
|     Addresses[Level] = Address | LShiftU64 (Indices[Level], Shifts[Level]);
 | |
| 
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Update the maximum address of freed page which can be used for memory
 | |
|   // promotion upon out-of-memory-space.
 | |
|   //
 | |
|   GetLastGuardedFreePageAddress (&Address);
 | |
|   if (Address != 0) {
 | |
|     mLastPromotedPage = Address;
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   This function checks to see if the given memory map descriptor in a memory map
 | |
|   can be merged with any guarded free pages.
 | |
| 
 | |
|   @param  MemoryMapEntry    A pointer to a descriptor in MemoryMap.
 | |
|   @param  MaxAddress        Maximum address to stop the merge.
 | |
| 
 | |
|   @return VOID
 | |
| 
 | |
| **/
 | |
| VOID
 | |
| MergeGuardPages (
 | |
|   IN EFI_MEMORY_DESCRIPTOR      *MemoryMapEntry,
 | |
|   IN EFI_PHYSICAL_ADDRESS       MaxAddress
 | |
|   )
 | |
| {
 | |
|   EFI_PHYSICAL_ADDRESS        EndAddress;
 | |
|   UINT64                      Bitmap;
 | |
|   INTN                        Pages;
 | |
| 
 | |
|   if (!IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED) ||
 | |
|       MemoryMapEntry->Type >= EfiMemoryMappedIO) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   Bitmap = 0;
 | |
|   Pages  = EFI_SIZE_TO_PAGES ((UINTN)(MaxAddress - MemoryMapEntry->PhysicalStart));
 | |
|   Pages -= (INTN)MemoryMapEntry->NumberOfPages;
 | |
|   while (Pages > 0) {
 | |
|     if (Bitmap == 0) {
 | |
|       EndAddress = MemoryMapEntry->PhysicalStart +
 | |
|                    EFI_PAGES_TO_SIZE ((UINTN)MemoryMapEntry->NumberOfPages);
 | |
|       Bitmap = GetGuardedMemoryBits (EndAddress, GUARDED_HEAP_MAP_ENTRY_BITS);
 | |
|     }
 | |
| 
 | |
|     if ((Bitmap & 1) == 0) {
 | |
|       break;
 | |
|     }
 | |
| 
 | |
|     Pages--;
 | |
|     MemoryMapEntry->NumberOfPages++;
 | |
|     Bitmap = RShiftU64 (Bitmap, 1);
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Put part (at most 64 pages a time) guarded free pages back to free page pool.
 | |
| 
 | |
|   Freed memory guard is used to detect Use-After-Free (UAF) memory issue, which
 | |
|   makes use of 'Used then throw away' way to detect any illegal access to freed
 | |
|   memory. The thrown-away memory will be marked as not-present so that any access
 | |
|   to those memory (after free) will be caught by page-fault exception.
 | |
| 
 | |
|   The problem is that this will consume lots of memory space. Once no memory
 | |
|   left in pool to allocate, we have to restore part of the freed pages to their
 | |
|   normal function. Otherwise the whole system will stop functioning.
 | |
| 
 | |
|   @param  StartAddress    Start address of promoted memory.
 | |
|   @param  EndAddress      End address of promoted memory.
 | |
| 
 | |
|   @return TRUE    Succeeded to promote memory.
 | |
|   @return FALSE   No free memory found.
 | |
| 
 | |
| **/
 | |
| BOOLEAN
 | |
| PromoteGuardedFreePages (
 | |
|   OUT EFI_PHYSICAL_ADDRESS      *StartAddress,
 | |
|   OUT EFI_PHYSICAL_ADDRESS      *EndAddress
 | |
|   )
 | |
| {
 | |
|   EFI_STATUS              Status;
 | |
|   UINTN                   AvailablePages;
 | |
|   UINT64                  Bitmap;
 | |
|   EFI_PHYSICAL_ADDRESS    Start;
 | |
| 
 | |
|   if (!IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED)) {
 | |
|     return FALSE;
 | |
|   }
 | |
| 
 | |
|   //
 | |
|   // Similar to memory allocation service, always search the freed pages in
 | |
|   // descending direction.
 | |
|   //
 | |
|   Start           = mLastPromotedPage;
 | |
|   AvailablePages  = 0;
 | |
|   while (AvailablePages == 0) {
 | |
|     Start -= EFI_PAGES_TO_SIZE (GUARDED_HEAP_MAP_ENTRY_BITS);
 | |
|     //
 | |
|     // If the address wraps around, try the really freed pages at top.
 | |
|     //
 | |
|     if (Start > mLastPromotedPage) {
 | |
|       GetLastGuardedFreePageAddress (&Start);
 | |
|       ASSERT (Start != 0);
 | |
|       Start -= EFI_PAGES_TO_SIZE (GUARDED_HEAP_MAP_ENTRY_BITS);
 | |
|     }
 | |
| 
 | |
|     Bitmap = GetGuardedMemoryBits (Start, GUARDED_HEAP_MAP_ENTRY_BITS);
 | |
|     while (Bitmap > 0) {
 | |
|       if ((Bitmap & 1) != 0) {
 | |
|         ++AvailablePages;
 | |
|       } else if (AvailablePages == 0) {
 | |
|         Start += EFI_PAGES_TO_SIZE (1);
 | |
|       } else {
 | |
|         break;
 | |
|       }
 | |
| 
 | |
|       Bitmap = RShiftU64 (Bitmap, 1);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   if (AvailablePages != 0) {
 | |
|     DEBUG ((DEBUG_INFO, "Promoted pages: %lX (%lx)\r\n", Start, (UINT64)AvailablePages));
 | |
|     ClearGuardedMemoryBits (Start, AvailablePages);
 | |
| 
 | |
|     if (gCpu != NULL) {
 | |
|       //
 | |
|       // Set flag to make sure allocating memory without GUARD for page table
 | |
|       // operation; otherwise infinite loops could be caused.
 | |
|       //
 | |
|       mOnGuarding = TRUE;
 | |
|       Status = gCpu->SetMemoryAttributes (gCpu, Start, EFI_PAGES_TO_SIZE(AvailablePages), 0);
 | |
|       ASSERT_EFI_ERROR (Status);
 | |
|       mOnGuarding = FALSE;
 | |
|     }
 | |
| 
 | |
|     mLastPromotedPage = Start;
 | |
|     *StartAddress     = Start;
 | |
|     *EndAddress       = Start + EFI_PAGES_TO_SIZE (AvailablePages) - 1;
 | |
|     return TRUE;
 | |
|   }
 | |
| 
 | |
|   return FALSE;
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Notify function used to set all Guard pages before CPU Arch Protocol installed.
 | |
| **/
 | |
| VOID
 | |
| HeapGuardCpuArchProtocolNotify (
 | |
|   VOID
 | |
|   )
 | |
| {
 | |
|   ASSERT (gCpu != NULL);
 | |
| 
 | |
|   if (IsHeapGuardEnabled (GUARD_HEAP_TYPE_PAGE|GUARD_HEAP_TYPE_POOL) &&
 | |
|       IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED)) {
 | |
|     DEBUG ((DEBUG_ERROR, "Heap guard and freed memory guard cannot be enabled at the same time.\n"));
 | |
|     CpuDeadLoop ();
 | |
|   }
 | |
| 
 | |
|   if (IsHeapGuardEnabled (GUARD_HEAP_TYPE_PAGE|GUARD_HEAP_TYPE_POOL)) {
 | |
|     SetAllGuardPages ();
 | |
|   }
 | |
| 
 | |
|   if (IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED)) {
 | |
|     GuardAllFreedPages ();
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Helper function to convert a UINT64 value in binary to a string.
 | |
| 
 | |
|   @param[in]  Value       Value of a UINT64 integer.
 | |
|   @param[out]  BinString   String buffer to contain the conversion result.
 | |
| 
 | |
|   @return VOID.
 | |
| **/
 | |
| VOID
 | |
| Uint64ToBinString (
 | |
|   IN  UINT64      Value,
 | |
|   OUT CHAR8       *BinString
 | |
|   )
 | |
| {
 | |
|   UINTN Index;
 | |
| 
 | |
|   if (BinString == NULL) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   for (Index = 64; Index > 0; --Index) {
 | |
|     BinString[Index - 1] = '0' + (Value & 1);
 | |
|     Value = RShiftU64 (Value, 1);
 | |
|   }
 | |
|   BinString[64] = '\0';
 | |
| }
 | |
| 
 | |
| /**
 | |
|   Dump the guarded memory bit map.
 | |
| **/
 | |
| VOID
 | |
| EFIAPI
 | |
| DumpGuardedMemoryBitmap (
 | |
|   VOID
 | |
|   )
 | |
| {
 | |
|   UINTN     Entries[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINTN     Shifts[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINTN     Indices[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINT64    Tables[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINT64    Addresses[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | |
|   UINT64    TableEntry;
 | |
|   UINT64    Address;
 | |
|   INTN      Level;
 | |
|   UINTN     RepeatZero;
 | |
|   CHAR8     String[GUARDED_HEAP_MAP_ENTRY_BITS + 1];
 | |
|   CHAR8     *Ruler1;
 | |
|   CHAR8     *Ruler2;
 | |
| 
 | |
|   if (!IsHeapGuardEnabled (GUARD_HEAP_TYPE_ALL)) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   if (mGuardedMemoryMap == 0 ||
 | |
|       mMapLevel == 0 ||
 | |
|       mMapLevel > GUARDED_HEAP_MAP_TABLE_DEPTH) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   Ruler1 = "               3               2               1               0";
 | |
|   Ruler2 = "FEDCBA9876543210FEDCBA9876543210FEDCBA9876543210FEDCBA9876543210";
 | |
| 
 | |
|   DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "============================="
 | |
|                                   " Guarded Memory Bitmap "
 | |
|                                   "==============================\r\n"));
 | |
|   DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "                  %a\r\n", Ruler1));
 | |
|   DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "                  %a\r\n", Ruler2));
 | |
| 
 | |
|   CopyMem (Entries, mLevelMask, sizeof (Entries));
 | |
|   CopyMem (Shifts, mLevelShift, sizeof (Shifts));
 | |
| 
 | |
|   SetMem (Indices, sizeof(Indices), 0);
 | |
|   SetMem (Tables, sizeof(Tables), 0);
 | |
|   SetMem (Addresses, sizeof(Addresses), 0);
 | |
| 
 | |
|   Level         = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | |
|   Tables[Level] = mGuardedMemoryMap;
 | |
|   Address       = 0;
 | |
|   RepeatZero    = 0;
 | |
| 
 | |
|   while (TRUE) {
 | |
|     if (Indices[Level] > Entries[Level]) {
 | |
| 
 | |
|       Tables[Level] = 0;
 | |
|       Level        -= 1;
 | |
|       RepeatZero    = 0;
 | |
| 
 | |
|       DEBUG ((
 | |
|         HEAP_GUARD_DEBUG_LEVEL,
 | |
|         "========================================="
 | |
|         "=========================================\r\n"
 | |
|         ));
 | |
| 
 | |
|     } else {
 | |
| 
 | |
|       TableEntry  = ((UINT64 *)(UINTN)Tables[Level])[Indices[Level]];
 | |
|       Address     = Addresses[Level];
 | |
| 
 | |
|       if (TableEntry == 0) {
 | |
| 
 | |
|         if (Level == GUARDED_HEAP_MAP_TABLE_DEPTH - 1) {
 | |
|           if (RepeatZero == 0) {
 | |
|             Uint64ToBinString(TableEntry, String);
 | |
|             DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "%016lx: %a\r\n", Address, String));
 | |
|           } else if (RepeatZero == 1) {
 | |
|             DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "...             : ...\r\n"));
 | |
|           }
 | |
|           RepeatZero += 1;
 | |
|         }
 | |
| 
 | |
|       } else if (Level < GUARDED_HEAP_MAP_TABLE_DEPTH - 1) {
 | |
| 
 | |
|         Level            += 1;
 | |
|         Tables[Level]     = TableEntry;
 | |
|         Addresses[Level]  = Address;
 | |
|         Indices[Level]    = 0;
 | |
|         RepeatZero        = 0;
 | |
| 
 | |
|         continue;
 | |
| 
 | |
|       } else {
 | |
| 
 | |
|         RepeatZero = 0;
 | |
|         Uint64ToBinString(TableEntry, String);
 | |
|         DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "%016lx: %a\r\n", Address, String));
 | |
| 
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     if (Level < (GUARDED_HEAP_MAP_TABLE_DEPTH - (INTN)mMapLevel)) {
 | |
|       break;
 | |
|     }
 | |
| 
 | |
|     Indices[Level] += 1;
 | |
|     Address = (Level == 0) ? 0 : Addresses[Level - 1];
 | |
|     Addresses[Level] = Address | LShiftU64(Indices[Level], Shifts[Level]);
 | |
| 
 | |
|   }
 | |
| }
 | |
| 
 |