MapMemory is not initialized by FindGuardedMemoryMap or CoreInternalAllocatePages which calls MapMemory. So we give a 0 to it. Cc: Dandan Bi <dandan.bi@intel.com> Cc: Liming Gao <liming.gao@intel.com> Signed-off-by: Shenglei Zhang <shenglei.zhang@intel.com> Reviewed-by: Dandan Bi <dandan.bi@intel.com>
		
			
				
	
	
		
			1747 lines
		
	
	
		
			46 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1747 lines
		
	
	
		
			46 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/** @file
 | 
						|
  UEFI Heap Guard functions.
 | 
						|
 | 
						|
Copyright (c) 2017-2018, Intel Corporation. All rights reserved.<BR>
 | 
						|
SPDX-License-Identifier: BSD-2-Clause-Patent
 | 
						|
 | 
						|
**/
 | 
						|
 | 
						|
#include "DxeMain.h"
 | 
						|
#include "Imem.h"
 | 
						|
#include "HeapGuard.h"
 | 
						|
 | 
						|
//
 | 
						|
// Global to avoid infinite reentrance of memory allocation when updating
 | 
						|
// page table attributes, which may need allocate pages for new PDE/PTE.
 | 
						|
//
 | 
						|
GLOBAL_REMOVE_IF_UNREFERENCED BOOLEAN mOnGuarding = FALSE;
 | 
						|
 | 
						|
//
 | 
						|
// Pointer to table tracking the Guarded memory with bitmap, in which  '1'
 | 
						|
// is used to indicate memory guarded. '0' might be free memory or Guard
 | 
						|
// page itself, depending on status of memory adjacent to it.
 | 
						|
//
 | 
						|
GLOBAL_REMOVE_IF_UNREFERENCED UINT64 mGuardedMemoryMap = 0;
 | 
						|
 | 
						|
//
 | 
						|
// Current depth level of map table pointed by mGuardedMemoryMap.
 | 
						|
// mMapLevel must be initialized at least by 1. It will be automatically
 | 
						|
// updated according to the address of memory just tracked.
 | 
						|
//
 | 
						|
GLOBAL_REMOVE_IF_UNREFERENCED UINTN mMapLevel = 1;
 | 
						|
 | 
						|
//
 | 
						|
// Shift and mask for each level of map table
 | 
						|
//
 | 
						|
GLOBAL_REMOVE_IF_UNREFERENCED UINTN mLevelShift[GUARDED_HEAP_MAP_TABLE_DEPTH]
 | 
						|
                                    = GUARDED_HEAP_MAP_TABLE_DEPTH_SHIFTS;
 | 
						|
GLOBAL_REMOVE_IF_UNREFERENCED UINTN mLevelMask[GUARDED_HEAP_MAP_TABLE_DEPTH]
 | 
						|
                                    = GUARDED_HEAP_MAP_TABLE_DEPTH_MASKS;
 | 
						|
 | 
						|
//
 | 
						|
// Used for promoting freed but not used pages.
 | 
						|
//
 | 
						|
GLOBAL_REMOVE_IF_UNREFERENCED EFI_PHYSICAL_ADDRESS mLastPromotedPage = BASE_4GB;
 | 
						|
 | 
						|
/**
 | 
						|
  Set corresponding bits in bitmap table to 1 according to the address.
 | 
						|
 | 
						|
  @param[in]  Address     Start address to set for.
 | 
						|
  @param[in]  BitNumber   Number of bits to set.
 | 
						|
  @param[in]  BitMap      Pointer to bitmap which covers the Address.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
STATIC
 | 
						|
VOID
 | 
						|
SetBits (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Address,
 | 
						|
  IN UINTN                   BitNumber,
 | 
						|
  IN UINT64                  *BitMap
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINTN           Lsbs;
 | 
						|
  UINTN           Qwords;
 | 
						|
  UINTN           Msbs;
 | 
						|
  UINTN           StartBit;
 | 
						|
  UINTN           EndBit;
 | 
						|
 | 
						|
  StartBit  = (UINTN)GUARDED_HEAP_MAP_ENTRY_BIT_INDEX (Address);
 | 
						|
  EndBit    = (StartBit + BitNumber - 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
 | 
						|
  if ((StartBit + BitNumber) >= GUARDED_HEAP_MAP_ENTRY_BITS) {
 | 
						|
    Msbs    = (GUARDED_HEAP_MAP_ENTRY_BITS - StartBit) %
 | 
						|
              GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
    Lsbs    = (EndBit + 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
    Qwords  = (BitNumber - Msbs) / GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
  } else {
 | 
						|
    Msbs    = BitNumber;
 | 
						|
    Lsbs    = 0;
 | 
						|
    Qwords  = 0;
 | 
						|
  }
 | 
						|
 | 
						|
  if (Msbs > 0) {
 | 
						|
    *BitMap |= LShiftU64 (LShiftU64 (1, Msbs) - 1, StartBit);
 | 
						|
    BitMap  += 1;
 | 
						|
  }
 | 
						|
 | 
						|
  if (Qwords > 0) {
 | 
						|
    SetMem64 ((VOID *)BitMap, Qwords * GUARDED_HEAP_MAP_ENTRY_BYTES,
 | 
						|
              (UINT64)-1);
 | 
						|
    BitMap += Qwords;
 | 
						|
  }
 | 
						|
 | 
						|
  if (Lsbs > 0) {
 | 
						|
    *BitMap |= (LShiftU64 (1, Lsbs) - 1);
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Set corresponding bits in bitmap table to 0 according to the address.
 | 
						|
 | 
						|
  @param[in]  Address     Start address to set for.
 | 
						|
  @param[in]  BitNumber   Number of bits to set.
 | 
						|
  @param[in]  BitMap      Pointer to bitmap which covers the Address.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
STATIC
 | 
						|
VOID
 | 
						|
ClearBits (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Address,
 | 
						|
  IN UINTN                   BitNumber,
 | 
						|
  IN UINT64                  *BitMap
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINTN           Lsbs;
 | 
						|
  UINTN           Qwords;
 | 
						|
  UINTN           Msbs;
 | 
						|
  UINTN           StartBit;
 | 
						|
  UINTN           EndBit;
 | 
						|
 | 
						|
  StartBit  = (UINTN)GUARDED_HEAP_MAP_ENTRY_BIT_INDEX (Address);
 | 
						|
  EndBit    = (StartBit + BitNumber - 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
 | 
						|
  if ((StartBit + BitNumber) >= GUARDED_HEAP_MAP_ENTRY_BITS) {
 | 
						|
    Msbs    = (GUARDED_HEAP_MAP_ENTRY_BITS - StartBit) %
 | 
						|
              GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
    Lsbs    = (EndBit + 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
    Qwords  = (BitNumber - Msbs) / GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
  } else {
 | 
						|
    Msbs    = BitNumber;
 | 
						|
    Lsbs    = 0;
 | 
						|
    Qwords  = 0;
 | 
						|
  }
 | 
						|
 | 
						|
  if (Msbs > 0) {
 | 
						|
    *BitMap &= ~LShiftU64 (LShiftU64 (1, Msbs) - 1, StartBit);
 | 
						|
    BitMap  += 1;
 | 
						|
  }
 | 
						|
 | 
						|
  if (Qwords > 0) {
 | 
						|
    SetMem64 ((VOID *)BitMap, Qwords * GUARDED_HEAP_MAP_ENTRY_BYTES, 0);
 | 
						|
    BitMap += Qwords;
 | 
						|
  }
 | 
						|
 | 
						|
  if (Lsbs > 0) {
 | 
						|
    *BitMap &= ~(LShiftU64 (1, Lsbs) - 1);
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Get corresponding bits in bitmap table according to the address.
 | 
						|
 | 
						|
  The value of bit 0 corresponds to the status of memory at given Address.
 | 
						|
  No more than 64 bits can be retrieved in one call.
 | 
						|
 | 
						|
  @param[in]  Address     Start address to retrieve bits for.
 | 
						|
  @param[in]  BitNumber   Number of bits to get.
 | 
						|
  @param[in]  BitMap      Pointer to bitmap which covers the Address.
 | 
						|
 | 
						|
  @return An integer containing the bits information.
 | 
						|
**/
 | 
						|
STATIC
 | 
						|
UINT64
 | 
						|
GetBits (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Address,
 | 
						|
  IN UINTN                   BitNumber,
 | 
						|
  IN UINT64                  *BitMap
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINTN           StartBit;
 | 
						|
  UINTN           EndBit;
 | 
						|
  UINTN           Lsbs;
 | 
						|
  UINTN           Msbs;
 | 
						|
  UINT64          Result;
 | 
						|
 | 
						|
  ASSERT (BitNumber <= GUARDED_HEAP_MAP_ENTRY_BITS);
 | 
						|
 | 
						|
  StartBit  = (UINTN)GUARDED_HEAP_MAP_ENTRY_BIT_INDEX (Address);
 | 
						|
  EndBit    = (StartBit + BitNumber - 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
 | 
						|
  if ((StartBit + BitNumber) > GUARDED_HEAP_MAP_ENTRY_BITS) {
 | 
						|
    Msbs = GUARDED_HEAP_MAP_ENTRY_BITS - StartBit;
 | 
						|
    Lsbs = (EndBit + 1) % GUARDED_HEAP_MAP_ENTRY_BITS;
 | 
						|
  } else {
 | 
						|
    Msbs = BitNumber;
 | 
						|
    Lsbs = 0;
 | 
						|
  }
 | 
						|
 | 
						|
  if (StartBit == 0 && BitNumber == GUARDED_HEAP_MAP_ENTRY_BITS) {
 | 
						|
    Result = *BitMap;
 | 
						|
  } else {
 | 
						|
    Result    = RShiftU64((*BitMap), StartBit) & (LShiftU64(1, Msbs) - 1);
 | 
						|
    if (Lsbs > 0) {
 | 
						|
      BitMap  += 1;
 | 
						|
      Result  |= LShiftU64 ((*BitMap) & (LShiftU64 (1, Lsbs) - 1), Msbs);
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  return Result;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Locate the pointer of bitmap from the guarded memory bitmap tables, which
 | 
						|
  covers the given Address.
 | 
						|
 | 
						|
  @param[in]  Address       Start address to search the bitmap for.
 | 
						|
  @param[in]  AllocMapUnit  Flag to indicate memory allocation for the table.
 | 
						|
  @param[out] BitMap        Pointer to bitmap which covers the Address.
 | 
						|
 | 
						|
  @return The bit number from given Address to the end of current map table.
 | 
						|
**/
 | 
						|
UINTN
 | 
						|
FindGuardedMemoryMap (
 | 
						|
  IN  EFI_PHYSICAL_ADDRESS    Address,
 | 
						|
  IN  BOOLEAN                 AllocMapUnit,
 | 
						|
  OUT UINT64                  **BitMap
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINTN                   Level;
 | 
						|
  UINT64                  *GuardMap;
 | 
						|
  UINT64                  MapMemory;
 | 
						|
  UINTN                   Index;
 | 
						|
  UINTN                   Size;
 | 
						|
  UINTN                   BitsToUnitEnd;
 | 
						|
  EFI_STATUS              Status;
 | 
						|
 | 
						|
  MapMemory = 0;
 | 
						|
 | 
						|
  //
 | 
						|
  // Adjust current map table depth according to the address to access
 | 
						|
  //
 | 
						|
  while (AllocMapUnit &&
 | 
						|
         mMapLevel < GUARDED_HEAP_MAP_TABLE_DEPTH &&
 | 
						|
         RShiftU64 (
 | 
						|
           Address,
 | 
						|
           mLevelShift[GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel - 1]
 | 
						|
           ) != 0) {
 | 
						|
 | 
						|
    if (mGuardedMemoryMap != 0) {
 | 
						|
      Size = (mLevelMask[GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel - 1] + 1)
 | 
						|
             * GUARDED_HEAP_MAP_ENTRY_BYTES;
 | 
						|
      Status = CoreInternalAllocatePages (
 | 
						|
                  AllocateAnyPages,
 | 
						|
                  EfiBootServicesData,
 | 
						|
                  EFI_SIZE_TO_PAGES (Size),
 | 
						|
                  &MapMemory,
 | 
						|
                  FALSE
 | 
						|
                  );
 | 
						|
      ASSERT_EFI_ERROR (Status);
 | 
						|
      ASSERT (MapMemory != 0);
 | 
						|
 | 
						|
      SetMem ((VOID *)(UINTN)MapMemory, Size, 0);
 | 
						|
 | 
						|
      *(UINT64 *)(UINTN)MapMemory = mGuardedMemoryMap;
 | 
						|
      mGuardedMemoryMap = MapMemory;
 | 
						|
    }
 | 
						|
 | 
						|
    mMapLevel++;
 | 
						|
 | 
						|
  }
 | 
						|
 | 
						|
  GuardMap = &mGuardedMemoryMap;
 | 
						|
  for (Level = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | 
						|
       Level < GUARDED_HEAP_MAP_TABLE_DEPTH;
 | 
						|
       ++Level) {
 | 
						|
 | 
						|
    if (*GuardMap == 0) {
 | 
						|
      if (!AllocMapUnit) {
 | 
						|
        GuardMap = NULL;
 | 
						|
        break;
 | 
						|
      }
 | 
						|
 | 
						|
      Size = (mLevelMask[Level] + 1) * GUARDED_HEAP_MAP_ENTRY_BYTES;
 | 
						|
      Status = CoreInternalAllocatePages (
 | 
						|
                  AllocateAnyPages,
 | 
						|
                  EfiBootServicesData,
 | 
						|
                  EFI_SIZE_TO_PAGES (Size),
 | 
						|
                  &MapMemory,
 | 
						|
                  FALSE
 | 
						|
                  );
 | 
						|
      ASSERT_EFI_ERROR (Status);
 | 
						|
      ASSERT (MapMemory != 0);
 | 
						|
 | 
						|
      SetMem ((VOID *)(UINTN)MapMemory, Size, 0);
 | 
						|
      *GuardMap = MapMemory;
 | 
						|
    }
 | 
						|
 | 
						|
    Index     = (UINTN)RShiftU64 (Address, mLevelShift[Level]);
 | 
						|
    Index     &= mLevelMask[Level];
 | 
						|
    GuardMap  = (UINT64 *)(UINTN)((*GuardMap) + Index * sizeof (UINT64));
 | 
						|
 | 
						|
  }
 | 
						|
 | 
						|
  BitsToUnitEnd = GUARDED_HEAP_MAP_BITS - GUARDED_HEAP_MAP_BIT_INDEX (Address);
 | 
						|
  *BitMap       = GuardMap;
 | 
						|
 | 
						|
  return BitsToUnitEnd;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Set corresponding bits in bitmap table to 1 according to given memory range.
 | 
						|
 | 
						|
  @param[in]  Address       Memory address to guard from.
 | 
						|
  @param[in]  NumberOfPages Number of pages to guard.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
EFIAPI
 | 
						|
SetGuardedMemoryBits (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Address,
 | 
						|
  IN UINTN                   NumberOfPages
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINT64            *BitMap;
 | 
						|
  UINTN             Bits;
 | 
						|
  UINTN             BitsToUnitEnd;
 | 
						|
 | 
						|
  while (NumberOfPages > 0) {
 | 
						|
    BitsToUnitEnd = FindGuardedMemoryMap (Address, TRUE, &BitMap);
 | 
						|
    ASSERT (BitMap != NULL);
 | 
						|
 | 
						|
    if (NumberOfPages > BitsToUnitEnd) {
 | 
						|
      // Cross map unit
 | 
						|
      Bits = BitsToUnitEnd;
 | 
						|
    } else {
 | 
						|
      Bits  = NumberOfPages;
 | 
						|
    }
 | 
						|
 | 
						|
    SetBits (Address, Bits, BitMap);
 | 
						|
 | 
						|
    NumberOfPages -= Bits;
 | 
						|
    Address       += EFI_PAGES_TO_SIZE (Bits);
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Clear corresponding bits in bitmap table according to given memory range.
 | 
						|
 | 
						|
  @param[in]  Address       Memory address to unset from.
 | 
						|
  @param[in]  NumberOfPages Number of pages to unset guard.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
EFIAPI
 | 
						|
ClearGuardedMemoryBits (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Address,
 | 
						|
  IN UINTN                   NumberOfPages
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINT64            *BitMap;
 | 
						|
  UINTN             Bits;
 | 
						|
  UINTN             BitsToUnitEnd;
 | 
						|
 | 
						|
  while (NumberOfPages > 0) {
 | 
						|
    BitsToUnitEnd = FindGuardedMemoryMap (Address, TRUE, &BitMap);
 | 
						|
    ASSERT (BitMap != NULL);
 | 
						|
 | 
						|
    if (NumberOfPages > BitsToUnitEnd) {
 | 
						|
      // Cross map unit
 | 
						|
      Bits = BitsToUnitEnd;
 | 
						|
    } else {
 | 
						|
      Bits  = NumberOfPages;
 | 
						|
    }
 | 
						|
 | 
						|
    ClearBits (Address, Bits, BitMap);
 | 
						|
 | 
						|
    NumberOfPages -= Bits;
 | 
						|
    Address       += EFI_PAGES_TO_SIZE (Bits);
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Retrieve corresponding bits in bitmap table according to given memory range.
 | 
						|
 | 
						|
  @param[in]  Address       Memory address to retrieve from.
 | 
						|
  @param[in]  NumberOfPages Number of pages to retrieve.
 | 
						|
 | 
						|
  @return An integer containing the guarded memory bitmap.
 | 
						|
**/
 | 
						|
UINT64
 | 
						|
GetGuardedMemoryBits (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Address,
 | 
						|
  IN UINTN                   NumberOfPages
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINT64            *BitMap;
 | 
						|
  UINTN             Bits;
 | 
						|
  UINT64            Result;
 | 
						|
  UINTN             Shift;
 | 
						|
  UINTN             BitsToUnitEnd;
 | 
						|
 | 
						|
  ASSERT (NumberOfPages <= GUARDED_HEAP_MAP_ENTRY_BITS);
 | 
						|
 | 
						|
  Result = 0;
 | 
						|
  Shift  = 0;
 | 
						|
  while (NumberOfPages > 0) {
 | 
						|
    BitsToUnitEnd = FindGuardedMemoryMap (Address, FALSE, &BitMap);
 | 
						|
 | 
						|
    if (NumberOfPages > BitsToUnitEnd) {
 | 
						|
      // Cross map unit
 | 
						|
      Bits  = BitsToUnitEnd;
 | 
						|
    } else {
 | 
						|
      Bits  = NumberOfPages;
 | 
						|
    }
 | 
						|
 | 
						|
    if (BitMap != NULL) {
 | 
						|
      Result |= LShiftU64 (GetBits (Address, Bits, BitMap), Shift);
 | 
						|
    }
 | 
						|
 | 
						|
    Shift         += Bits;
 | 
						|
    NumberOfPages -= Bits;
 | 
						|
    Address       += EFI_PAGES_TO_SIZE (Bits);
 | 
						|
  }
 | 
						|
 | 
						|
  return Result;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Get bit value in bitmap table for the given address.
 | 
						|
 | 
						|
  @param[in]  Address     The address to retrieve for.
 | 
						|
 | 
						|
  @return 1 or 0.
 | 
						|
**/
 | 
						|
UINTN
 | 
						|
EFIAPI
 | 
						|
GetGuardMapBit (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Address
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINT64        *GuardMap;
 | 
						|
 | 
						|
  FindGuardedMemoryMap (Address, FALSE, &GuardMap);
 | 
						|
  if (GuardMap != NULL) {
 | 
						|
    if (RShiftU64 (*GuardMap,
 | 
						|
                   GUARDED_HEAP_MAP_ENTRY_BIT_INDEX (Address)) & 1) {
 | 
						|
      return 1;
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  return 0;
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
/**
 | 
						|
  Check to see if the page at the given address is a Guard page or not.
 | 
						|
 | 
						|
  @param[in]  Address     The address to check for.
 | 
						|
 | 
						|
  @return TRUE  The page at Address is a Guard page.
 | 
						|
  @return FALSE The page at Address is not a Guard page.
 | 
						|
**/
 | 
						|
BOOLEAN
 | 
						|
EFIAPI
 | 
						|
IsGuardPage (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Address
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINT64        BitMap;
 | 
						|
 | 
						|
  //
 | 
						|
  // There must be at least one guarded page before and/or after given
 | 
						|
  // address if it's a Guard page. The bitmap pattern should be one of
 | 
						|
  // 001, 100 and 101
 | 
						|
  //
 | 
						|
  BitMap = GetGuardedMemoryBits (Address - EFI_PAGE_SIZE, 3);
 | 
						|
  return ((BitMap == BIT0) || (BitMap == BIT2) || (BitMap == (BIT2 | BIT0)));
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
/**
 | 
						|
  Check to see if the page at the given address is guarded or not.
 | 
						|
 | 
						|
  @param[in]  Address     The address to check for.
 | 
						|
 | 
						|
  @return TRUE  The page at Address is guarded.
 | 
						|
  @return FALSE The page at Address is not guarded.
 | 
						|
**/
 | 
						|
BOOLEAN
 | 
						|
EFIAPI
 | 
						|
IsMemoryGuarded (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Address
 | 
						|
  )
 | 
						|
{
 | 
						|
  return (GetGuardMapBit (Address) == 1);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Set the page at the given address to be a Guard page.
 | 
						|
 | 
						|
  This is done by changing the page table attribute to be NOT PRSENT.
 | 
						|
 | 
						|
  @param[in]  BaseAddress     Page address to Guard at
 | 
						|
 | 
						|
  @return VOID
 | 
						|
**/
 | 
						|
VOID
 | 
						|
EFIAPI
 | 
						|
SetGuardPage (
 | 
						|
  IN  EFI_PHYSICAL_ADDRESS      BaseAddress
 | 
						|
  )
 | 
						|
{
 | 
						|
  EFI_STATUS      Status;
 | 
						|
 | 
						|
  if (gCpu == NULL) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Set flag to make sure allocating memory without GUARD for page table
 | 
						|
  // operation; otherwise infinite loops could be caused.
 | 
						|
  //
 | 
						|
  mOnGuarding = TRUE;
 | 
						|
  //
 | 
						|
  // Note: This might overwrite other attributes needed by other features,
 | 
						|
  // such as NX memory protection.
 | 
						|
  //
 | 
						|
  Status = gCpu->SetMemoryAttributes (gCpu, BaseAddress, EFI_PAGE_SIZE, EFI_MEMORY_RP);
 | 
						|
  ASSERT_EFI_ERROR (Status);
 | 
						|
  mOnGuarding = FALSE;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Unset the Guard page at the given address to the normal memory.
 | 
						|
 | 
						|
  This is done by changing the page table attribute to be PRSENT.
 | 
						|
 | 
						|
  @param[in]  BaseAddress     Page address to Guard at.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
EFIAPI
 | 
						|
UnsetGuardPage (
 | 
						|
  IN  EFI_PHYSICAL_ADDRESS      BaseAddress
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINT64          Attributes;
 | 
						|
  EFI_STATUS      Status;
 | 
						|
 | 
						|
  if (gCpu == NULL) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Once the Guard page is unset, it will be freed back to memory pool. NX
 | 
						|
  // memory protection must be restored for this page if NX is enabled for free
 | 
						|
  // memory.
 | 
						|
  //
 | 
						|
  Attributes = 0;
 | 
						|
  if ((PcdGet64 (PcdDxeNxMemoryProtectionPolicy) & (1 << EfiConventionalMemory)) != 0) {
 | 
						|
    Attributes |= EFI_MEMORY_XP;
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Set flag to make sure allocating memory without GUARD for page table
 | 
						|
  // operation; otherwise infinite loops could be caused.
 | 
						|
  //
 | 
						|
  mOnGuarding = TRUE;
 | 
						|
  //
 | 
						|
  // Note: This might overwrite other attributes needed by other features,
 | 
						|
  // such as memory protection (NX). Please make sure they are not enabled
 | 
						|
  // at the same time.
 | 
						|
  //
 | 
						|
  Status = gCpu->SetMemoryAttributes (gCpu, BaseAddress, EFI_PAGE_SIZE, Attributes);
 | 
						|
  ASSERT_EFI_ERROR (Status);
 | 
						|
  mOnGuarding = FALSE;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Check to see if the memory at the given address should be guarded or not.
 | 
						|
 | 
						|
  @param[in]  MemoryType      Memory type to check.
 | 
						|
  @param[in]  AllocateType    Allocation type to check.
 | 
						|
  @param[in]  PageOrPool      Indicate a page allocation or pool allocation.
 | 
						|
 | 
						|
 | 
						|
  @return TRUE  The given type of memory should be guarded.
 | 
						|
  @return FALSE The given type of memory should not be guarded.
 | 
						|
**/
 | 
						|
BOOLEAN
 | 
						|
IsMemoryTypeToGuard (
 | 
						|
  IN EFI_MEMORY_TYPE        MemoryType,
 | 
						|
  IN EFI_ALLOCATE_TYPE      AllocateType,
 | 
						|
  IN UINT8                  PageOrPool
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINT64 TestBit;
 | 
						|
  UINT64 ConfigBit;
 | 
						|
 | 
						|
  if (AllocateType == AllocateAddress) {
 | 
						|
    return FALSE;
 | 
						|
  }
 | 
						|
 | 
						|
  if ((PcdGet8 (PcdHeapGuardPropertyMask) & PageOrPool) == 0) {
 | 
						|
    return FALSE;
 | 
						|
  }
 | 
						|
 | 
						|
  if (PageOrPool == GUARD_HEAP_TYPE_POOL) {
 | 
						|
    ConfigBit = PcdGet64 (PcdHeapGuardPoolType);
 | 
						|
  } else if (PageOrPool == GUARD_HEAP_TYPE_PAGE) {
 | 
						|
    ConfigBit = PcdGet64 (PcdHeapGuardPageType);
 | 
						|
  } else {
 | 
						|
    ConfigBit = (UINT64)-1;
 | 
						|
  }
 | 
						|
 | 
						|
  if ((UINT32)MemoryType >= MEMORY_TYPE_OS_RESERVED_MIN) {
 | 
						|
    TestBit = BIT63;
 | 
						|
  } else if ((UINT32) MemoryType >= MEMORY_TYPE_OEM_RESERVED_MIN) {
 | 
						|
    TestBit = BIT62;
 | 
						|
  } else if (MemoryType < EfiMaxMemoryType) {
 | 
						|
    TestBit = LShiftU64 (1, MemoryType);
 | 
						|
  } else if (MemoryType == EfiMaxMemoryType) {
 | 
						|
    TestBit = (UINT64)-1;
 | 
						|
  } else {
 | 
						|
    TestBit = 0;
 | 
						|
  }
 | 
						|
 | 
						|
  return ((ConfigBit & TestBit) != 0);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Check to see if the pool at the given address should be guarded or not.
 | 
						|
 | 
						|
  @param[in]  MemoryType      Pool type to check.
 | 
						|
 | 
						|
 | 
						|
  @return TRUE  The given type of pool should be guarded.
 | 
						|
  @return FALSE The given type of pool should not be guarded.
 | 
						|
**/
 | 
						|
BOOLEAN
 | 
						|
IsPoolTypeToGuard (
 | 
						|
  IN EFI_MEMORY_TYPE        MemoryType
 | 
						|
  )
 | 
						|
{
 | 
						|
  return IsMemoryTypeToGuard (MemoryType, AllocateAnyPages,
 | 
						|
                              GUARD_HEAP_TYPE_POOL);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Check to see if the page at the given address should be guarded or not.
 | 
						|
 | 
						|
  @param[in]  MemoryType      Page type to check.
 | 
						|
  @param[in]  AllocateType    Allocation type to check.
 | 
						|
 | 
						|
  @return TRUE  The given type of page should be guarded.
 | 
						|
  @return FALSE The given type of page should not be guarded.
 | 
						|
**/
 | 
						|
BOOLEAN
 | 
						|
IsPageTypeToGuard (
 | 
						|
  IN EFI_MEMORY_TYPE        MemoryType,
 | 
						|
  IN EFI_ALLOCATE_TYPE      AllocateType
 | 
						|
  )
 | 
						|
{
 | 
						|
  return IsMemoryTypeToGuard (MemoryType, AllocateType, GUARD_HEAP_TYPE_PAGE);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Check to see if the heap guard is enabled for page and/or pool allocation.
 | 
						|
 | 
						|
  @param[in]  GuardType   Specify the sub-type(s) of Heap Guard.
 | 
						|
 | 
						|
  @return TRUE/FALSE.
 | 
						|
**/
 | 
						|
BOOLEAN
 | 
						|
IsHeapGuardEnabled (
 | 
						|
  UINT8           GuardType
 | 
						|
  )
 | 
						|
{
 | 
						|
  return IsMemoryTypeToGuard (EfiMaxMemoryType, AllocateAnyPages, GuardType);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Set head Guard and tail Guard for the given memory range.
 | 
						|
 | 
						|
  @param[in]  Memory          Base address of memory to set guard for.
 | 
						|
  @param[in]  NumberOfPages   Memory size in pages.
 | 
						|
 | 
						|
  @return VOID
 | 
						|
**/
 | 
						|
VOID
 | 
						|
SetGuardForMemory (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS   Memory,
 | 
						|
  IN UINTN                  NumberOfPages
 | 
						|
  )
 | 
						|
{
 | 
						|
  EFI_PHYSICAL_ADDRESS    GuardPage;
 | 
						|
 | 
						|
  //
 | 
						|
  // Set tail Guard
 | 
						|
  //
 | 
						|
  GuardPage = Memory + EFI_PAGES_TO_SIZE (NumberOfPages);
 | 
						|
  if (!IsGuardPage (GuardPage)) {
 | 
						|
    SetGuardPage (GuardPage);
 | 
						|
  }
 | 
						|
 | 
						|
  // Set head Guard
 | 
						|
  GuardPage = Memory - EFI_PAGES_TO_SIZE (1);
 | 
						|
  if (!IsGuardPage (GuardPage)) {
 | 
						|
    SetGuardPage (GuardPage);
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Mark the memory range as Guarded
 | 
						|
  //
 | 
						|
  SetGuardedMemoryBits (Memory, NumberOfPages);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Unset head Guard and tail Guard for the given memory range.
 | 
						|
 | 
						|
  @param[in]  Memory          Base address of memory to unset guard for.
 | 
						|
  @param[in]  NumberOfPages   Memory size in pages.
 | 
						|
 | 
						|
  @return VOID
 | 
						|
**/
 | 
						|
VOID
 | 
						|
UnsetGuardForMemory (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS   Memory,
 | 
						|
  IN UINTN                  NumberOfPages
 | 
						|
  )
 | 
						|
{
 | 
						|
  EFI_PHYSICAL_ADDRESS  GuardPage;
 | 
						|
  UINT64                GuardBitmap;
 | 
						|
 | 
						|
  if (NumberOfPages == 0) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Head Guard must be one page before, if any.
 | 
						|
  //
 | 
						|
  //          MSB-> 1     0 <-LSB
 | 
						|
  //          -------------------
 | 
						|
  //  Head Guard -> 0     1 -> Don't free Head Guard  (shared Guard)
 | 
						|
  //  Head Guard -> 0     0 -> Free Head Guard either (not shared Guard)
 | 
						|
  //                1     X -> Don't free first page  (need a new Guard)
 | 
						|
  //                           (it'll be turned into a Guard page later)
 | 
						|
  //          -------------------
 | 
						|
  //      Start -> -1    -2
 | 
						|
  //
 | 
						|
  GuardPage = Memory - EFI_PAGES_TO_SIZE (1);
 | 
						|
  GuardBitmap = GetGuardedMemoryBits (Memory - EFI_PAGES_TO_SIZE (2), 2);
 | 
						|
  if ((GuardBitmap & BIT1) == 0) {
 | 
						|
    //
 | 
						|
    // Head Guard exists.
 | 
						|
    //
 | 
						|
    if ((GuardBitmap & BIT0) == 0) {
 | 
						|
      //
 | 
						|
      // If the head Guard is not a tail Guard of adjacent memory block,
 | 
						|
      // unset it.
 | 
						|
      //
 | 
						|
      UnsetGuardPage (GuardPage);
 | 
						|
    }
 | 
						|
  } else {
 | 
						|
    //
 | 
						|
    // Pages before memory to free are still in Guard. It's a partial free
 | 
						|
    // case. Turn first page of memory block to free into a new Guard.
 | 
						|
    //
 | 
						|
    SetGuardPage (Memory);
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Tail Guard must be the page after this memory block to free, if any.
 | 
						|
  //
 | 
						|
  //   MSB-> 1     0 <-LSB
 | 
						|
  //  --------------------
 | 
						|
  //         1     0 <- Tail Guard -> Don't free Tail Guard  (shared Guard)
 | 
						|
  //         0     0 <- Tail Guard -> Free Tail Guard either (not shared Guard)
 | 
						|
  //         X     1               -> Don't free last page   (need a new Guard)
 | 
						|
  //                                 (it'll be turned into a Guard page later)
 | 
						|
  //  --------------------
 | 
						|
  //        +1    +0 <- End
 | 
						|
  //
 | 
						|
  GuardPage = Memory + EFI_PAGES_TO_SIZE (NumberOfPages);
 | 
						|
  GuardBitmap = GetGuardedMemoryBits (GuardPage, 2);
 | 
						|
  if ((GuardBitmap & BIT0) == 0) {
 | 
						|
    //
 | 
						|
    // Tail Guard exists.
 | 
						|
    //
 | 
						|
    if ((GuardBitmap & BIT1) == 0) {
 | 
						|
      //
 | 
						|
      // If the tail Guard is not a head Guard of adjacent memory block,
 | 
						|
      // free it; otherwise, keep it.
 | 
						|
      //
 | 
						|
      UnsetGuardPage (GuardPage);
 | 
						|
    }
 | 
						|
  } else {
 | 
						|
    //
 | 
						|
    // Pages after memory to free are still in Guard. It's a partial free
 | 
						|
    // case. We need to keep one page to be a head Guard.
 | 
						|
    //
 | 
						|
    SetGuardPage (GuardPage - EFI_PAGES_TO_SIZE (1));
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // No matter what, we just clear the mark of the Guarded memory.
 | 
						|
  //
 | 
						|
  ClearGuardedMemoryBits(Memory, NumberOfPages);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Adjust address of free memory according to existing and/or required Guard.
 | 
						|
 | 
						|
  This function will check if there're existing Guard pages of adjacent
 | 
						|
  memory blocks, and try to use it as the Guard page of the memory to be
 | 
						|
  allocated.
 | 
						|
 | 
						|
  @param[in]  Start           Start address of free memory block.
 | 
						|
  @param[in]  Size            Size of free memory block.
 | 
						|
  @param[in]  SizeRequested   Size of memory to allocate.
 | 
						|
 | 
						|
  @return The end address of memory block found.
 | 
						|
  @return 0 if no enough space for the required size of memory and its Guard.
 | 
						|
**/
 | 
						|
UINT64
 | 
						|
AdjustMemoryS (
 | 
						|
  IN UINT64                  Start,
 | 
						|
  IN UINT64                  Size,
 | 
						|
  IN UINT64                  SizeRequested
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINT64  Target;
 | 
						|
 | 
						|
  //
 | 
						|
  // UEFI spec requires that allocated pool must be 8-byte aligned. If it's
 | 
						|
  // indicated to put the pool near the Tail Guard, we need extra bytes to
 | 
						|
  // make sure alignment of the returned pool address.
 | 
						|
  //
 | 
						|
  if ((PcdGet8 (PcdHeapGuardPropertyMask) & BIT7) == 0) {
 | 
						|
    SizeRequested = ALIGN_VALUE(SizeRequested, 8);
 | 
						|
  }
 | 
						|
 | 
						|
  Target = Start + Size - SizeRequested;
 | 
						|
  ASSERT (Target >= Start);
 | 
						|
  if (Target == 0) {
 | 
						|
    return 0;
 | 
						|
  }
 | 
						|
 | 
						|
  if (!IsGuardPage (Start + Size)) {
 | 
						|
    // No Guard at tail to share. One more page is needed.
 | 
						|
    Target -= EFI_PAGES_TO_SIZE (1);
 | 
						|
  }
 | 
						|
 | 
						|
  // Out of range?
 | 
						|
  if (Target < Start) {
 | 
						|
    return 0;
 | 
						|
  }
 | 
						|
 | 
						|
  // At the edge?
 | 
						|
  if (Target == Start) {
 | 
						|
    if (!IsGuardPage (Target - EFI_PAGES_TO_SIZE (1))) {
 | 
						|
      // No enough space for a new head Guard if no Guard at head to share.
 | 
						|
      return 0;
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  // OK, we have enough pages for memory and its Guards. Return the End of the
 | 
						|
  // free space.
 | 
						|
  return Target + SizeRequested - 1;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Adjust the start address and number of pages to free according to Guard.
 | 
						|
 | 
						|
  The purpose of this function is to keep the shared Guard page with adjacent
 | 
						|
  memory block if it's still in guard, or free it if no more sharing. Another
 | 
						|
  is to reserve pages as Guard pages in partial page free situation.
 | 
						|
 | 
						|
  @param[in,out]  Memory          Base address of memory to free.
 | 
						|
  @param[in,out]  NumberOfPages   Size of memory to free.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
AdjustMemoryF (
 | 
						|
  IN OUT EFI_PHYSICAL_ADDRESS    *Memory,
 | 
						|
  IN OUT UINTN                   *NumberOfPages
 | 
						|
  )
 | 
						|
{
 | 
						|
  EFI_PHYSICAL_ADDRESS  Start;
 | 
						|
  EFI_PHYSICAL_ADDRESS  MemoryToTest;
 | 
						|
  UINTN                 PagesToFree;
 | 
						|
  UINT64                GuardBitmap;
 | 
						|
 | 
						|
  if (Memory == NULL || NumberOfPages == NULL || *NumberOfPages == 0) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  Start = *Memory;
 | 
						|
  PagesToFree = *NumberOfPages;
 | 
						|
 | 
						|
  //
 | 
						|
  // Head Guard must be one page before, if any.
 | 
						|
  //
 | 
						|
  //          MSB-> 1     0 <-LSB
 | 
						|
  //          -------------------
 | 
						|
  //  Head Guard -> 0     1 -> Don't free Head Guard  (shared Guard)
 | 
						|
  //  Head Guard -> 0     0 -> Free Head Guard either (not shared Guard)
 | 
						|
  //                1     X -> Don't free first page  (need a new Guard)
 | 
						|
  //                           (it'll be turned into a Guard page later)
 | 
						|
  //          -------------------
 | 
						|
  //      Start -> -1    -2
 | 
						|
  //
 | 
						|
  MemoryToTest = Start - EFI_PAGES_TO_SIZE (2);
 | 
						|
  GuardBitmap = GetGuardedMemoryBits (MemoryToTest, 2);
 | 
						|
  if ((GuardBitmap & BIT1) == 0) {
 | 
						|
    //
 | 
						|
    // Head Guard exists.
 | 
						|
    //
 | 
						|
    if ((GuardBitmap & BIT0) == 0) {
 | 
						|
      //
 | 
						|
      // If the head Guard is not a tail Guard of adjacent memory block,
 | 
						|
      // free it; otherwise, keep it.
 | 
						|
      //
 | 
						|
      Start       -= EFI_PAGES_TO_SIZE (1);
 | 
						|
      PagesToFree += 1;
 | 
						|
    }
 | 
						|
  } else {
 | 
						|
    //
 | 
						|
    // No Head Guard, and pages before memory to free are still in Guard. It's a
 | 
						|
    // partial free case. We need to keep one page to be a tail Guard.
 | 
						|
    //
 | 
						|
    Start       += EFI_PAGES_TO_SIZE (1);
 | 
						|
    PagesToFree -= 1;
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Tail Guard must be the page after this memory block to free, if any.
 | 
						|
  //
 | 
						|
  //   MSB-> 1     0 <-LSB
 | 
						|
  //  --------------------
 | 
						|
  //         1     0 <- Tail Guard -> Don't free Tail Guard  (shared Guard)
 | 
						|
  //         0     0 <- Tail Guard -> Free Tail Guard either (not shared Guard)
 | 
						|
  //         X     1               -> Don't free last page   (need a new Guard)
 | 
						|
  //                                 (it'll be turned into a Guard page later)
 | 
						|
  //  --------------------
 | 
						|
  //        +1    +0 <- End
 | 
						|
  //
 | 
						|
  MemoryToTest = Start + EFI_PAGES_TO_SIZE (PagesToFree);
 | 
						|
  GuardBitmap = GetGuardedMemoryBits (MemoryToTest, 2);
 | 
						|
  if ((GuardBitmap & BIT0) == 0) {
 | 
						|
    //
 | 
						|
    // Tail Guard exists.
 | 
						|
    //
 | 
						|
    if ((GuardBitmap & BIT1) == 0) {
 | 
						|
      //
 | 
						|
      // If the tail Guard is not a head Guard of adjacent memory block,
 | 
						|
      // free it; otherwise, keep it.
 | 
						|
      //
 | 
						|
      PagesToFree += 1;
 | 
						|
    }
 | 
						|
  } else if (PagesToFree > 0) {
 | 
						|
    //
 | 
						|
    // No Tail Guard, and pages after memory to free are still in Guard. It's a
 | 
						|
    // partial free case. We need to keep one page to be a head Guard.
 | 
						|
    //
 | 
						|
    PagesToFree -= 1;
 | 
						|
  }
 | 
						|
 | 
						|
  *Memory         = Start;
 | 
						|
  *NumberOfPages  = PagesToFree;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Adjust the base and number of pages to really allocate according to Guard.
 | 
						|
 | 
						|
  @param[in,out]  Memory          Base address of free memory.
 | 
						|
  @param[in,out]  NumberOfPages   Size of memory to allocate.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
AdjustMemoryA (
 | 
						|
  IN OUT EFI_PHYSICAL_ADDRESS    *Memory,
 | 
						|
  IN OUT UINTN                   *NumberOfPages
 | 
						|
  )
 | 
						|
{
 | 
						|
  //
 | 
						|
  // FindFreePages() has already taken the Guard into account. It's safe to
 | 
						|
  // adjust the start address and/or number of pages here, to make sure that
 | 
						|
  // the Guards are also "allocated".
 | 
						|
  //
 | 
						|
  if (!IsGuardPage (*Memory + EFI_PAGES_TO_SIZE (*NumberOfPages))) {
 | 
						|
    // No tail Guard, add one.
 | 
						|
    *NumberOfPages += 1;
 | 
						|
  }
 | 
						|
 | 
						|
  if (!IsGuardPage (*Memory - EFI_PAGE_SIZE)) {
 | 
						|
    // No head Guard, add one.
 | 
						|
    *Memory        -= EFI_PAGE_SIZE;
 | 
						|
    *NumberOfPages += 1;
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Adjust the pool head position to make sure the Guard page is adjavent to
 | 
						|
  pool tail or pool head.
 | 
						|
 | 
						|
  @param[in]  Memory    Base address of memory allocated.
 | 
						|
  @param[in]  NoPages   Number of pages actually allocated.
 | 
						|
  @param[in]  Size      Size of memory requested.
 | 
						|
                        (plus pool head/tail overhead)
 | 
						|
 | 
						|
  @return Address of pool head.
 | 
						|
**/
 | 
						|
VOID *
 | 
						|
AdjustPoolHeadA (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Memory,
 | 
						|
  IN UINTN                   NoPages,
 | 
						|
  IN UINTN                   Size
 | 
						|
  )
 | 
						|
{
 | 
						|
  if (Memory == 0 || (PcdGet8 (PcdHeapGuardPropertyMask) & BIT7) != 0) {
 | 
						|
    //
 | 
						|
    // Pool head is put near the head Guard
 | 
						|
    //
 | 
						|
    return (VOID *)(UINTN)Memory;
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Pool head is put near the tail Guard
 | 
						|
  //
 | 
						|
  Size = ALIGN_VALUE (Size, 8);
 | 
						|
  return (VOID *)(UINTN)(Memory + EFI_PAGES_TO_SIZE (NoPages) - Size);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Get the page base address according to pool head address.
 | 
						|
 | 
						|
  @param[in]  Memory    Head address of pool to free.
 | 
						|
 | 
						|
  @return Address of pool head.
 | 
						|
**/
 | 
						|
VOID *
 | 
						|
AdjustPoolHeadF (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS    Memory
 | 
						|
  )
 | 
						|
{
 | 
						|
  if (Memory == 0 || (PcdGet8 (PcdHeapGuardPropertyMask) & BIT7) != 0) {
 | 
						|
    //
 | 
						|
    // Pool head is put near the head Guard
 | 
						|
    //
 | 
						|
    return (VOID *)(UINTN)Memory;
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Pool head is put near the tail Guard
 | 
						|
  //
 | 
						|
  return (VOID *)(UINTN)(Memory & ~EFI_PAGE_MASK);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Allocate or free guarded memory.
 | 
						|
 | 
						|
  @param[in]  Start           Start address of memory to allocate or free.
 | 
						|
  @param[in]  NumberOfPages   Memory size in pages.
 | 
						|
  @param[in]  NewType         Memory type to convert to.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
EFI_STATUS
 | 
						|
CoreConvertPagesWithGuard (
 | 
						|
  IN UINT64           Start,
 | 
						|
  IN UINTN            NumberOfPages,
 | 
						|
  IN EFI_MEMORY_TYPE  NewType
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINT64  OldStart;
 | 
						|
  UINTN   OldPages;
 | 
						|
 | 
						|
  if (NewType == EfiConventionalMemory) {
 | 
						|
    OldStart = Start;
 | 
						|
    OldPages = NumberOfPages;
 | 
						|
 | 
						|
    AdjustMemoryF (&Start, &NumberOfPages);
 | 
						|
    //
 | 
						|
    // It's safe to unset Guard page inside memory lock because there should
 | 
						|
    // be no memory allocation occurred in updating memory page attribute at
 | 
						|
    // this point. And unsetting Guard page before free will prevent Guard
 | 
						|
    // page just freed back to pool from being allocated right away before
 | 
						|
    // marking it usable (from non-present to present).
 | 
						|
    //
 | 
						|
    UnsetGuardForMemory (OldStart, OldPages);
 | 
						|
    if (NumberOfPages == 0) {
 | 
						|
      return EFI_SUCCESS;
 | 
						|
    }
 | 
						|
  } else {
 | 
						|
    AdjustMemoryA (&Start, &NumberOfPages);
 | 
						|
  }
 | 
						|
 | 
						|
  return CoreConvertPages (Start, NumberOfPages, NewType);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Set all Guard pages which cannot be set before CPU Arch Protocol installed.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
SetAllGuardPages (
 | 
						|
  VOID
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINTN     Entries[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINTN     Shifts[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINTN     Indices[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINT64    Tables[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINT64    Addresses[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINT64    TableEntry;
 | 
						|
  UINT64    Address;
 | 
						|
  UINT64    GuardPage;
 | 
						|
  INTN      Level;
 | 
						|
  UINTN     Index;
 | 
						|
  BOOLEAN   OnGuarding;
 | 
						|
 | 
						|
  if (mGuardedMemoryMap == 0 ||
 | 
						|
      mMapLevel == 0 ||
 | 
						|
      mMapLevel > GUARDED_HEAP_MAP_TABLE_DEPTH) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  CopyMem (Entries, mLevelMask, sizeof (Entries));
 | 
						|
  CopyMem (Shifts, mLevelShift, sizeof (Shifts));
 | 
						|
 | 
						|
  SetMem (Tables, sizeof(Tables), 0);
 | 
						|
  SetMem (Addresses, sizeof(Addresses), 0);
 | 
						|
  SetMem (Indices, sizeof(Indices), 0);
 | 
						|
 | 
						|
  Level         = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | 
						|
  Tables[Level] = mGuardedMemoryMap;
 | 
						|
  Address       = 0;
 | 
						|
  OnGuarding    = FALSE;
 | 
						|
 | 
						|
  DEBUG_CODE (
 | 
						|
    DumpGuardedMemoryBitmap ();
 | 
						|
  );
 | 
						|
 | 
						|
  while (TRUE) {
 | 
						|
    if (Indices[Level] > Entries[Level]) {
 | 
						|
      Tables[Level] = 0;
 | 
						|
      Level        -= 1;
 | 
						|
    } else {
 | 
						|
 | 
						|
      TableEntry  = ((UINT64 *)(UINTN)(Tables[Level]))[Indices[Level]];
 | 
						|
      Address     = Addresses[Level];
 | 
						|
 | 
						|
      if (TableEntry == 0) {
 | 
						|
 | 
						|
        OnGuarding = FALSE;
 | 
						|
 | 
						|
      } else if (Level < GUARDED_HEAP_MAP_TABLE_DEPTH - 1) {
 | 
						|
 | 
						|
        Level            += 1;
 | 
						|
        Tables[Level]     = TableEntry;
 | 
						|
        Addresses[Level]  = Address;
 | 
						|
        Indices[Level]    = 0;
 | 
						|
 | 
						|
        continue;
 | 
						|
 | 
						|
      } else {
 | 
						|
 | 
						|
        Index = 0;
 | 
						|
        while (Index < GUARDED_HEAP_MAP_ENTRY_BITS) {
 | 
						|
          if ((TableEntry & 1) == 1) {
 | 
						|
            if (OnGuarding) {
 | 
						|
              GuardPage = 0;
 | 
						|
            } else {
 | 
						|
              GuardPage = Address - EFI_PAGE_SIZE;
 | 
						|
            }
 | 
						|
            OnGuarding = TRUE;
 | 
						|
          } else {
 | 
						|
            if (OnGuarding) {
 | 
						|
              GuardPage = Address;
 | 
						|
            } else {
 | 
						|
              GuardPage = 0;
 | 
						|
            }
 | 
						|
            OnGuarding = FALSE;
 | 
						|
          }
 | 
						|
 | 
						|
          if (GuardPage != 0) {
 | 
						|
            SetGuardPage (GuardPage);
 | 
						|
          }
 | 
						|
 | 
						|
          if (TableEntry == 0) {
 | 
						|
            break;
 | 
						|
          }
 | 
						|
 | 
						|
          TableEntry = RShiftU64 (TableEntry, 1);
 | 
						|
          Address   += EFI_PAGE_SIZE;
 | 
						|
          Index     += 1;
 | 
						|
        }
 | 
						|
      }
 | 
						|
    }
 | 
						|
 | 
						|
    if (Level < (GUARDED_HEAP_MAP_TABLE_DEPTH - (INTN)mMapLevel)) {
 | 
						|
      break;
 | 
						|
    }
 | 
						|
 | 
						|
    Indices[Level] += 1;
 | 
						|
    Address = (Level == 0) ? 0 : Addresses[Level - 1];
 | 
						|
    Addresses[Level] = Address | LShiftU64(Indices[Level], Shifts[Level]);
 | 
						|
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Find the address of top-most guarded free page.
 | 
						|
 | 
						|
  @param[out]  Address    Start address of top-most guarded free page.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
GetLastGuardedFreePageAddress (
 | 
						|
  OUT EFI_PHYSICAL_ADDRESS      *Address
 | 
						|
  )
 | 
						|
{
 | 
						|
  EFI_PHYSICAL_ADDRESS    AddressGranularity;
 | 
						|
  EFI_PHYSICAL_ADDRESS    BaseAddress;
 | 
						|
  UINTN                   Level;
 | 
						|
  UINT64                  Map;
 | 
						|
  INTN                    Index;
 | 
						|
 | 
						|
  ASSERT (mMapLevel >= 1);
 | 
						|
 | 
						|
  BaseAddress = 0;
 | 
						|
  Map = mGuardedMemoryMap;
 | 
						|
  for (Level = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | 
						|
       Level < GUARDED_HEAP_MAP_TABLE_DEPTH;
 | 
						|
       ++Level) {
 | 
						|
    AddressGranularity = LShiftU64 (1, mLevelShift[Level]);
 | 
						|
 | 
						|
    //
 | 
						|
    // Find the non-NULL entry at largest index.
 | 
						|
    //
 | 
						|
    for (Index = (INTN)mLevelMask[Level]; Index >= 0 ; --Index) {
 | 
						|
      if (((UINT64 *)(UINTN)Map)[Index] != 0) {
 | 
						|
        BaseAddress += MultU64x32 (AddressGranularity, (UINT32)Index);
 | 
						|
        Map = ((UINT64 *)(UINTN)Map)[Index];
 | 
						|
        break;
 | 
						|
      }
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Find the non-zero MSB then get the page address.
 | 
						|
  //
 | 
						|
  while (Map != 0) {
 | 
						|
    Map = RShiftU64 (Map, 1);
 | 
						|
    BaseAddress += EFI_PAGES_TO_SIZE (1);
 | 
						|
  }
 | 
						|
 | 
						|
  *Address = BaseAddress;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Record freed pages.
 | 
						|
 | 
						|
  @param[in]  BaseAddress   Base address of just freed pages.
 | 
						|
  @param[in]  Pages         Number of freed pages.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
MarkFreedPages (
 | 
						|
  IN EFI_PHYSICAL_ADDRESS     BaseAddress,
 | 
						|
  IN UINTN                    Pages
 | 
						|
  )
 | 
						|
{
 | 
						|
  SetGuardedMemoryBits (BaseAddress, Pages);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Record freed pages as well as mark them as not-present.
 | 
						|
 | 
						|
  @param[in]  BaseAddress   Base address of just freed pages.
 | 
						|
  @param[in]  Pages         Number of freed pages.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
EFIAPI
 | 
						|
GuardFreedPages (
 | 
						|
  IN  EFI_PHYSICAL_ADDRESS    BaseAddress,
 | 
						|
  IN  UINTN                   Pages
 | 
						|
  )
 | 
						|
{
 | 
						|
  EFI_STATUS      Status;
 | 
						|
 | 
						|
  //
 | 
						|
  // Legacy memory lower than 1MB might be accessed with no allocation. Leave
 | 
						|
  // them alone.
 | 
						|
  //
 | 
						|
  if (BaseAddress < BASE_1MB) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  MarkFreedPages (BaseAddress, Pages);
 | 
						|
  if (gCpu != NULL) {
 | 
						|
    //
 | 
						|
    // Set flag to make sure allocating memory without GUARD for page table
 | 
						|
    // operation; otherwise infinite loops could be caused.
 | 
						|
    //
 | 
						|
    mOnGuarding = TRUE;
 | 
						|
    //
 | 
						|
    // Note: This might overwrite other attributes needed by other features,
 | 
						|
    // such as NX memory protection.
 | 
						|
    //
 | 
						|
    Status = gCpu->SetMemoryAttributes (
 | 
						|
                     gCpu,
 | 
						|
                     BaseAddress,
 | 
						|
                     EFI_PAGES_TO_SIZE (Pages),
 | 
						|
                     EFI_MEMORY_RP
 | 
						|
                     );
 | 
						|
    //
 | 
						|
    // Normally we should ASSERT the returned Status. But there might be memory
 | 
						|
    // alloc/free involved in SetMemoryAttributes(), which might fail this
 | 
						|
    // calling. It's rare case so it's OK to let a few tiny holes be not-guarded.
 | 
						|
    //
 | 
						|
    if (EFI_ERROR (Status)) {
 | 
						|
      DEBUG ((DEBUG_WARN, "Failed to guard freed pages: %p (%lu)\n", BaseAddress, (UINT64)Pages));
 | 
						|
    }
 | 
						|
    mOnGuarding = FALSE;
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Record freed pages as well as mark them as not-present, if enabled.
 | 
						|
 | 
						|
  @param[in]  BaseAddress   Base address of just freed pages.
 | 
						|
  @param[in]  Pages         Number of freed pages.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
EFIAPI
 | 
						|
GuardFreedPagesChecked (
 | 
						|
  IN  EFI_PHYSICAL_ADDRESS    BaseAddress,
 | 
						|
  IN  UINTN                   Pages
 | 
						|
  )
 | 
						|
{
 | 
						|
  if (IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED)) {
 | 
						|
    GuardFreedPages (BaseAddress, Pages);
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Mark all pages freed before CPU Arch Protocol as not-present.
 | 
						|
 | 
						|
**/
 | 
						|
VOID
 | 
						|
GuardAllFreedPages (
 | 
						|
  VOID
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINTN     Entries[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINTN     Shifts[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINTN     Indices[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINT64    Tables[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINT64    Addresses[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINT64    TableEntry;
 | 
						|
  UINT64    Address;
 | 
						|
  UINT64    GuardPage;
 | 
						|
  INTN      Level;
 | 
						|
  UINT64    BitIndex;
 | 
						|
  UINTN     GuardPageNumber;
 | 
						|
 | 
						|
  if (mGuardedMemoryMap == 0 ||
 | 
						|
      mMapLevel == 0 ||
 | 
						|
      mMapLevel > GUARDED_HEAP_MAP_TABLE_DEPTH) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  CopyMem (Entries, mLevelMask, sizeof (Entries));
 | 
						|
  CopyMem (Shifts, mLevelShift, sizeof (Shifts));
 | 
						|
 | 
						|
  SetMem (Tables, sizeof(Tables), 0);
 | 
						|
  SetMem (Addresses, sizeof(Addresses), 0);
 | 
						|
  SetMem (Indices, sizeof(Indices), 0);
 | 
						|
 | 
						|
  Level           = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | 
						|
  Tables[Level]   = mGuardedMemoryMap;
 | 
						|
  Address         = 0;
 | 
						|
  GuardPage       = (UINT64)-1;
 | 
						|
  GuardPageNumber = 0;
 | 
						|
 | 
						|
  while (TRUE) {
 | 
						|
    if (Indices[Level] > Entries[Level]) {
 | 
						|
      Tables[Level] = 0;
 | 
						|
      Level        -= 1;
 | 
						|
    } else {
 | 
						|
      TableEntry  = ((UINT64 *)(UINTN)(Tables[Level]))[Indices[Level]];
 | 
						|
      Address     = Addresses[Level];
 | 
						|
 | 
						|
      if (Level < GUARDED_HEAP_MAP_TABLE_DEPTH - 1) {
 | 
						|
        Level            += 1;
 | 
						|
        Tables[Level]     = TableEntry;
 | 
						|
        Addresses[Level]  = Address;
 | 
						|
        Indices[Level]    = 0;
 | 
						|
 | 
						|
        continue;
 | 
						|
      } else {
 | 
						|
        BitIndex = 1;
 | 
						|
        while (BitIndex != 0) {
 | 
						|
          if ((TableEntry & BitIndex) != 0) {
 | 
						|
            if (GuardPage == (UINT64)-1) {
 | 
						|
              GuardPage = Address;
 | 
						|
            }
 | 
						|
            ++GuardPageNumber;
 | 
						|
          } else if (GuardPageNumber > 0) {
 | 
						|
            GuardFreedPages (GuardPage, GuardPageNumber);
 | 
						|
            GuardPageNumber = 0;
 | 
						|
            GuardPage       = (UINT64)-1;
 | 
						|
          }
 | 
						|
 | 
						|
          if (TableEntry == 0) {
 | 
						|
            break;
 | 
						|
          }
 | 
						|
 | 
						|
          Address += EFI_PAGES_TO_SIZE (1);
 | 
						|
          BitIndex = LShiftU64 (BitIndex, 1);
 | 
						|
        }
 | 
						|
      }
 | 
						|
    }
 | 
						|
 | 
						|
    if (Level < (GUARDED_HEAP_MAP_TABLE_DEPTH - (INTN)mMapLevel)) {
 | 
						|
      break;
 | 
						|
    }
 | 
						|
 | 
						|
    Indices[Level] += 1;
 | 
						|
    Address = (Level == 0) ? 0 : Addresses[Level - 1];
 | 
						|
    Addresses[Level] = Address | LShiftU64 (Indices[Level], Shifts[Level]);
 | 
						|
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Update the maximum address of freed page which can be used for memory
 | 
						|
  // promotion upon out-of-memory-space.
 | 
						|
  //
 | 
						|
  GetLastGuardedFreePageAddress (&Address);
 | 
						|
  if (Address != 0) {
 | 
						|
    mLastPromotedPage = Address;
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  This function checks to see if the given memory map descriptor in a memory map
 | 
						|
  can be merged with any guarded free pages.
 | 
						|
 | 
						|
  @param  MemoryMapEntry    A pointer to a descriptor in MemoryMap.
 | 
						|
  @param  MaxAddress        Maximum address to stop the merge.
 | 
						|
 | 
						|
  @return VOID
 | 
						|
 | 
						|
**/
 | 
						|
VOID
 | 
						|
MergeGuardPages (
 | 
						|
  IN EFI_MEMORY_DESCRIPTOR      *MemoryMapEntry,
 | 
						|
  IN EFI_PHYSICAL_ADDRESS       MaxAddress
 | 
						|
  )
 | 
						|
{
 | 
						|
  EFI_PHYSICAL_ADDRESS        EndAddress;
 | 
						|
  UINT64                      Bitmap;
 | 
						|
  INTN                        Pages;
 | 
						|
 | 
						|
  if (!IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED) ||
 | 
						|
      MemoryMapEntry->Type >= EfiMemoryMappedIO) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  Bitmap = 0;
 | 
						|
  Pages  = EFI_SIZE_TO_PAGES ((UINTN)(MaxAddress - MemoryMapEntry->PhysicalStart));
 | 
						|
  Pages -= (INTN)MemoryMapEntry->NumberOfPages;
 | 
						|
  while (Pages > 0) {
 | 
						|
    if (Bitmap == 0) {
 | 
						|
      EndAddress = MemoryMapEntry->PhysicalStart +
 | 
						|
                   EFI_PAGES_TO_SIZE ((UINTN)MemoryMapEntry->NumberOfPages);
 | 
						|
      Bitmap = GetGuardedMemoryBits (EndAddress, GUARDED_HEAP_MAP_ENTRY_BITS);
 | 
						|
    }
 | 
						|
 | 
						|
    if ((Bitmap & 1) == 0) {
 | 
						|
      break;
 | 
						|
    }
 | 
						|
 | 
						|
    Pages--;
 | 
						|
    MemoryMapEntry->NumberOfPages++;
 | 
						|
    Bitmap = RShiftU64 (Bitmap, 1);
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Put part (at most 64 pages a time) guarded free pages back to free page pool.
 | 
						|
 | 
						|
  Freed memory guard is used to detect Use-After-Free (UAF) memory issue, which
 | 
						|
  makes use of 'Used then throw away' way to detect any illegal access to freed
 | 
						|
  memory. The thrown-away memory will be marked as not-present so that any access
 | 
						|
  to those memory (after free) will be caught by page-fault exception.
 | 
						|
 | 
						|
  The problem is that this will consume lots of memory space. Once no memory
 | 
						|
  left in pool to allocate, we have to restore part of the freed pages to their
 | 
						|
  normal function. Otherwise the whole system will stop functioning.
 | 
						|
 | 
						|
  @param  StartAddress    Start address of promoted memory.
 | 
						|
  @param  EndAddress      End address of promoted memory.
 | 
						|
 | 
						|
  @return TRUE    Succeeded to promote memory.
 | 
						|
  @return FALSE   No free memory found.
 | 
						|
 | 
						|
**/
 | 
						|
BOOLEAN
 | 
						|
PromoteGuardedFreePages (
 | 
						|
  OUT EFI_PHYSICAL_ADDRESS      *StartAddress,
 | 
						|
  OUT EFI_PHYSICAL_ADDRESS      *EndAddress
 | 
						|
  )
 | 
						|
{
 | 
						|
  EFI_STATUS              Status;
 | 
						|
  UINTN                   AvailablePages;
 | 
						|
  UINT64                  Bitmap;
 | 
						|
  EFI_PHYSICAL_ADDRESS    Start;
 | 
						|
 | 
						|
  if (!IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED)) {
 | 
						|
    return FALSE;
 | 
						|
  }
 | 
						|
 | 
						|
  //
 | 
						|
  // Similar to memory allocation service, always search the freed pages in
 | 
						|
  // descending direction.
 | 
						|
  //
 | 
						|
  Start           = mLastPromotedPage;
 | 
						|
  AvailablePages  = 0;
 | 
						|
  while (AvailablePages == 0) {
 | 
						|
    Start -= EFI_PAGES_TO_SIZE (GUARDED_HEAP_MAP_ENTRY_BITS);
 | 
						|
    //
 | 
						|
    // If the address wraps around, try the really freed pages at top.
 | 
						|
    //
 | 
						|
    if (Start > mLastPromotedPage) {
 | 
						|
      GetLastGuardedFreePageAddress (&Start);
 | 
						|
      ASSERT (Start != 0);
 | 
						|
      Start -= EFI_PAGES_TO_SIZE (GUARDED_HEAP_MAP_ENTRY_BITS);
 | 
						|
    }
 | 
						|
 | 
						|
    Bitmap = GetGuardedMemoryBits (Start, GUARDED_HEAP_MAP_ENTRY_BITS);
 | 
						|
    while (Bitmap > 0) {
 | 
						|
      if ((Bitmap & 1) != 0) {
 | 
						|
        ++AvailablePages;
 | 
						|
      } else if (AvailablePages == 0) {
 | 
						|
        Start += EFI_PAGES_TO_SIZE (1);
 | 
						|
      } else {
 | 
						|
        break;
 | 
						|
      }
 | 
						|
 | 
						|
      Bitmap = RShiftU64 (Bitmap, 1);
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  if (AvailablePages != 0) {
 | 
						|
    DEBUG ((DEBUG_INFO, "Promoted pages: %lX (%lx)\r\n", Start, (UINT64)AvailablePages));
 | 
						|
    ClearGuardedMemoryBits (Start, AvailablePages);
 | 
						|
 | 
						|
    if (gCpu != NULL) {
 | 
						|
      //
 | 
						|
      // Set flag to make sure allocating memory without GUARD for page table
 | 
						|
      // operation; otherwise infinite loops could be caused.
 | 
						|
      //
 | 
						|
      mOnGuarding = TRUE;
 | 
						|
      Status = gCpu->SetMemoryAttributes (gCpu, Start, EFI_PAGES_TO_SIZE(AvailablePages), 0);
 | 
						|
      ASSERT_EFI_ERROR (Status);
 | 
						|
      mOnGuarding = FALSE;
 | 
						|
    }
 | 
						|
 | 
						|
    mLastPromotedPage = Start;
 | 
						|
    *StartAddress     = Start;
 | 
						|
    *EndAddress       = Start + EFI_PAGES_TO_SIZE (AvailablePages) - 1;
 | 
						|
    return TRUE;
 | 
						|
  }
 | 
						|
 | 
						|
  return FALSE;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Notify function used to set all Guard pages before CPU Arch Protocol installed.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
HeapGuardCpuArchProtocolNotify (
 | 
						|
  VOID
 | 
						|
  )
 | 
						|
{
 | 
						|
  ASSERT (gCpu != NULL);
 | 
						|
 | 
						|
  if (IsHeapGuardEnabled (GUARD_HEAP_TYPE_PAGE|GUARD_HEAP_TYPE_POOL) &&
 | 
						|
      IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED)) {
 | 
						|
    DEBUG ((DEBUG_ERROR, "Heap guard and freed memory guard cannot be enabled at the same time.\n"));
 | 
						|
    CpuDeadLoop ();
 | 
						|
  }
 | 
						|
 | 
						|
  if (IsHeapGuardEnabled (GUARD_HEAP_TYPE_PAGE|GUARD_HEAP_TYPE_POOL)) {
 | 
						|
    SetAllGuardPages ();
 | 
						|
  }
 | 
						|
 | 
						|
  if (IsHeapGuardEnabled (GUARD_HEAP_TYPE_FREED)) {
 | 
						|
    GuardAllFreedPages ();
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Helper function to convert a UINT64 value in binary to a string.
 | 
						|
 | 
						|
  @param[in]  Value       Value of a UINT64 integer.
 | 
						|
  @param[out]  BinString   String buffer to contain the conversion result.
 | 
						|
 | 
						|
  @return VOID.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
Uint64ToBinString (
 | 
						|
  IN  UINT64      Value,
 | 
						|
  OUT CHAR8       *BinString
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINTN Index;
 | 
						|
 | 
						|
  if (BinString == NULL) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  for (Index = 64; Index > 0; --Index) {
 | 
						|
    BinString[Index - 1] = '0' + (Value & 1);
 | 
						|
    Value = RShiftU64 (Value, 1);
 | 
						|
  }
 | 
						|
  BinString[64] = '\0';
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
  Dump the guarded memory bit map.
 | 
						|
**/
 | 
						|
VOID
 | 
						|
EFIAPI
 | 
						|
DumpGuardedMemoryBitmap (
 | 
						|
  VOID
 | 
						|
  )
 | 
						|
{
 | 
						|
  UINTN     Entries[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINTN     Shifts[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINTN     Indices[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINT64    Tables[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINT64    Addresses[GUARDED_HEAP_MAP_TABLE_DEPTH];
 | 
						|
  UINT64    TableEntry;
 | 
						|
  UINT64    Address;
 | 
						|
  INTN      Level;
 | 
						|
  UINTN     RepeatZero;
 | 
						|
  CHAR8     String[GUARDED_HEAP_MAP_ENTRY_BITS + 1];
 | 
						|
  CHAR8     *Ruler1;
 | 
						|
  CHAR8     *Ruler2;
 | 
						|
 | 
						|
  if (!IsHeapGuardEnabled (GUARD_HEAP_TYPE_ALL)) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  if (mGuardedMemoryMap == 0 ||
 | 
						|
      mMapLevel == 0 ||
 | 
						|
      mMapLevel > GUARDED_HEAP_MAP_TABLE_DEPTH) {
 | 
						|
    return;
 | 
						|
  }
 | 
						|
 | 
						|
  Ruler1 = "               3               2               1               0";
 | 
						|
  Ruler2 = "FEDCBA9876543210FEDCBA9876543210FEDCBA9876543210FEDCBA9876543210";
 | 
						|
 | 
						|
  DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "============================="
 | 
						|
                                  " Guarded Memory Bitmap "
 | 
						|
                                  "==============================\r\n"));
 | 
						|
  DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "                  %a\r\n", Ruler1));
 | 
						|
  DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "                  %a\r\n", Ruler2));
 | 
						|
 | 
						|
  CopyMem (Entries, mLevelMask, sizeof (Entries));
 | 
						|
  CopyMem (Shifts, mLevelShift, sizeof (Shifts));
 | 
						|
 | 
						|
  SetMem (Indices, sizeof(Indices), 0);
 | 
						|
  SetMem (Tables, sizeof(Tables), 0);
 | 
						|
  SetMem (Addresses, sizeof(Addresses), 0);
 | 
						|
 | 
						|
  Level         = GUARDED_HEAP_MAP_TABLE_DEPTH - mMapLevel;
 | 
						|
  Tables[Level] = mGuardedMemoryMap;
 | 
						|
  Address       = 0;
 | 
						|
  RepeatZero    = 0;
 | 
						|
 | 
						|
  while (TRUE) {
 | 
						|
    if (Indices[Level] > Entries[Level]) {
 | 
						|
 | 
						|
      Tables[Level] = 0;
 | 
						|
      Level        -= 1;
 | 
						|
      RepeatZero    = 0;
 | 
						|
 | 
						|
      DEBUG ((
 | 
						|
        HEAP_GUARD_DEBUG_LEVEL,
 | 
						|
        "========================================="
 | 
						|
        "=========================================\r\n"
 | 
						|
        ));
 | 
						|
 | 
						|
    } else {
 | 
						|
 | 
						|
      TableEntry  = ((UINT64 *)(UINTN)Tables[Level])[Indices[Level]];
 | 
						|
      Address     = Addresses[Level];
 | 
						|
 | 
						|
      if (TableEntry == 0) {
 | 
						|
 | 
						|
        if (Level == GUARDED_HEAP_MAP_TABLE_DEPTH - 1) {
 | 
						|
          if (RepeatZero == 0) {
 | 
						|
            Uint64ToBinString(TableEntry, String);
 | 
						|
            DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "%016lx: %a\r\n", Address, String));
 | 
						|
          } else if (RepeatZero == 1) {
 | 
						|
            DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "...             : ...\r\n"));
 | 
						|
          }
 | 
						|
          RepeatZero += 1;
 | 
						|
        }
 | 
						|
 | 
						|
      } else if (Level < GUARDED_HEAP_MAP_TABLE_DEPTH - 1) {
 | 
						|
 | 
						|
        Level            += 1;
 | 
						|
        Tables[Level]     = TableEntry;
 | 
						|
        Addresses[Level]  = Address;
 | 
						|
        Indices[Level]    = 0;
 | 
						|
        RepeatZero        = 0;
 | 
						|
 | 
						|
        continue;
 | 
						|
 | 
						|
      } else {
 | 
						|
 | 
						|
        RepeatZero = 0;
 | 
						|
        Uint64ToBinString(TableEntry, String);
 | 
						|
        DEBUG ((HEAP_GUARD_DEBUG_LEVEL, "%016lx: %a\r\n", Address, String));
 | 
						|
 | 
						|
      }
 | 
						|
    }
 | 
						|
 | 
						|
    if (Level < (GUARDED_HEAP_MAP_TABLE_DEPTH - (INTN)mMapLevel)) {
 | 
						|
      break;
 | 
						|
    }
 | 
						|
 | 
						|
    Indices[Level] += 1;
 | 
						|
    Address = (Level == 0) ? 0 : Addresses[Level - 1];
 | 
						|
    Addresses[Level] = Address | LShiftU64(Indices[Level], Shifts[Level]);
 | 
						|
 | 
						|
  }
 | 
						|
}
 | 
						|
 |