Provide get_mmu_ranges() for ARM64 to let payloads could get MMU ranges for all used memory regions. BUG=b:171858277 TEST=Build in x86, arm, arm64. emerge-zork libpayload depthcharge emerge-nyan libpayload depthcharge emerge-asurada libpayload depthcharge Signed-off-by: Meng-Huan Yu <menghuan@google.com> Change-Id: I39b24aefc9dbe530169b272e839d0e1e7c697742 Reviewed-on: https://review.coreboot.org/c/coreboot/+/48113 Tested-by: build bot (Jenkins) <no-reply@coreboot.org> Reviewed-by: Julius Werner <jwerner@chromium.org> Reviewed-by: Yu-Ping Wu <yupingso@google.com>
		
			
				
	
	
		
			206 lines
		
	
	
		
			7.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			206 lines
		
	
	
		
			7.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 *
 | 
						|
 * Copyright 2014 Google Inc.
 | 
						|
 *
 | 
						|
 * Redistribution and use in source and binary forms, with or without
 | 
						|
 * modification, are permitted provided that the following conditions
 | 
						|
 * are met:
 | 
						|
 * 1. Redistributions of source code must retain the above copyright
 | 
						|
 *    notice, this list of conditions and the following disclaimer.
 | 
						|
 * 2. Redistributions in binary form must reproduce the above copyright
 | 
						|
 *    notice, this list of conditions and the following disclaimer in the
 | 
						|
 *    documentation and/or other materials provided with the distribution.
 | 
						|
 * 3. The name of the author may not be used to endorse or promote products
 | 
						|
 *    derived from this software without specific prior written permission.
 | 
						|
 *
 | 
						|
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
 | 
						|
 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 | 
						|
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 | 
						|
 * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
 | 
						|
 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 | 
						|
 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
 | 
						|
 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 | 
						|
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 | 
						|
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 | 
						|
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 | 
						|
 * SUCH DAMAGE.
 | 
						|
 */
 | 
						|
 | 
						|
#ifndef __ARCH_ARM64_MMU_H__
 | 
						|
#define __ARCH_ARM64_MMU_H__
 | 
						|
 | 
						|
#include <libpayload.h>
 | 
						|
 | 
						|
struct mmu_memrange {
 | 
						|
	uint64_t base;
 | 
						|
	uint64_t size;
 | 
						|
	uint64_t type;
 | 
						|
};
 | 
						|
 | 
						|
struct mmu_ranges {
 | 
						|
	struct mmu_memrange entries[SYSINFO_MAX_MEM_RANGES];
 | 
						|
	size_t used;
 | 
						|
};
 | 
						|
 | 
						|
/*
 | 
						|
 * Symbols taken from linker script
 | 
						|
 * They mark the start and end of the region used by payload
 | 
						|
 */
 | 
						|
extern char _start[], _end[];
 | 
						|
 | 
						|
/* Memory attributes for mmap regions
 | 
						|
 * These attributes act as tag values for memrange regions
 | 
						|
 */
 | 
						|
 | 
						|
#define TYPE_NORMAL_MEM            1
 | 
						|
#define TYPE_DEV_MEM               2
 | 
						|
#define TYPE_DMA_MEM               3
 | 
						|
 | 
						|
/* Descriptor attributes */
 | 
						|
 | 
						|
#define INVALID_DESC               0x0
 | 
						|
#define BLOCK_DESC                 0x1
 | 
						|
#define TABLE_DESC                 0x3
 | 
						|
#define PAGE_DESC                  0x3
 | 
						|
 | 
						|
/* Block descriptor */
 | 
						|
#define BLOCK_NS                   (1 << 5)
 | 
						|
 | 
						|
#define BLOCK_AP_RW                (0 << 7)
 | 
						|
#define BLOCK_AP_RO                (1 << 7)
 | 
						|
 | 
						|
#define BLOCK_ACCESS               (1 << 10)
 | 
						|
 | 
						|
#define BLOCK_XN                   (1UL << 54)
 | 
						|
 | 
						|
#define BLOCK_SH_SHIFT                 (8)
 | 
						|
#define BLOCK_SH_NON_SHAREABLE         (0 << BLOCK_SH_SHIFT)
 | 
						|
#define BLOCK_SH_UNPREDICTABLE         (1 << BLOCK_SH_SHIFT)
 | 
						|
#define BLOCK_SH_OUTER_SHAREABLE       (2 << BLOCK_SH_SHIFT)
 | 
						|
#define BLOCK_SH_INNER_SHAREABLE       (3 << BLOCK_SH_SHIFT)
 | 
						|
 | 
						|
/* XLAT Table Init Attributes */
 | 
						|
 | 
						|
#define VA_START                   0x0
 | 
						|
#define BITS_PER_VA                48
 | 
						|
#define MIN_64_BIT_ADDR            (1UL << 32)
 | 
						|
/* Granule size of 4KB is being used */
 | 
						|
#define GRANULE_SIZE_SHIFT         12
 | 
						|
#define GRANULE_SIZE               (1 << GRANULE_SIZE_SHIFT)
 | 
						|
#define XLAT_TABLE_MASK            (~(0UL) << GRANULE_SIZE_SHIFT)
 | 
						|
#define GRANULE_SIZE_MASK          ((1 << GRANULE_SIZE_SHIFT) - 1)
 | 
						|
 | 
						|
#define BITS_RESOLVED_PER_LVL   (GRANULE_SIZE_SHIFT - 3)
 | 
						|
#define L0_ADDR_SHIFT           (GRANULE_SIZE_SHIFT + BITS_RESOLVED_PER_LVL * 3)
 | 
						|
#define L1_ADDR_SHIFT           (GRANULE_SIZE_SHIFT + BITS_RESOLVED_PER_LVL * 2)
 | 
						|
#define L2_ADDR_SHIFT           (GRANULE_SIZE_SHIFT + BITS_RESOLVED_PER_LVL * 1)
 | 
						|
#define L3_ADDR_SHIFT           (GRANULE_SIZE_SHIFT + BITS_RESOLVED_PER_LVL * 0)
 | 
						|
 | 
						|
#define L0_ADDR_MASK     (((1UL << BITS_RESOLVED_PER_LVL) - 1) << L0_ADDR_SHIFT)
 | 
						|
#define L1_ADDR_MASK     (((1UL << BITS_RESOLVED_PER_LVL) - 1) << L1_ADDR_SHIFT)
 | 
						|
#define L2_ADDR_MASK     (((1UL << BITS_RESOLVED_PER_LVL) - 1) << L2_ADDR_SHIFT)
 | 
						|
#define L3_ADDR_MASK     (((1UL << BITS_RESOLVED_PER_LVL) - 1) << L3_ADDR_SHIFT)
 | 
						|
 | 
						|
/* These macros give the size of the region addressed by each entry of a xlat
 | 
						|
   table at any given level */
 | 
						|
#define L3_XLAT_SIZE               (1UL << L3_ADDR_SHIFT)
 | 
						|
#define L2_XLAT_SIZE               (1UL << L2_ADDR_SHIFT)
 | 
						|
#define L1_XLAT_SIZE               (1UL << L1_ADDR_SHIFT)
 | 
						|
#define L0_XLAT_SIZE               (1UL << L0_ADDR_SHIFT)
 | 
						|
 | 
						|
/* Block indices required for MAIR */
 | 
						|
#define BLOCK_INDEX_MEM_DEV_NGNRNE 0
 | 
						|
#define BLOCK_INDEX_MEM_DEV_NGNRE  1
 | 
						|
#define BLOCK_INDEX_MEM_DEV_GRE    2
 | 
						|
#define BLOCK_INDEX_MEM_NORMAL_NC  3
 | 
						|
#define BLOCK_INDEX_MEM_NORMAL     4
 | 
						|
 | 
						|
#define BLOCK_INDEX_SHIFT          2
 | 
						|
 | 
						|
/* MAIR attributes */
 | 
						|
#define MAIR_ATTRIBUTES            ((0x00 << (BLOCK_INDEX_MEM_DEV_NGNRNE*8)) | \
 | 
						|
				    (0x04 << (BLOCK_INDEX_MEM_DEV_NGNRE*8))  | \
 | 
						|
				    (0x0c << (BLOCK_INDEX_MEM_DEV_GRE*8))    | \
 | 
						|
				    (0x44 << (BLOCK_INDEX_MEM_NORMAL_NC*8))  | \
 | 
						|
				    (0xffUL << (BLOCK_INDEX_MEM_NORMAL*8)))
 | 
						|
 | 
						|
/* TCR attributes */
 | 
						|
#define TCR_TOSZ                   (64 - BITS_PER_VA)
 | 
						|
 | 
						|
#define TCR_IRGN0_SHIFT            8
 | 
						|
#define TCR_IRGN0_NM_NC            (0x00 << TCR_IRGN0_SHIFT)
 | 
						|
#define TCR_IRGN0_NM_WBWAC         (0x01 << TCR_IRGN0_SHIFT)
 | 
						|
#define TCR_IRGN0_NM_WTC           (0x02 << TCR_IRGN0_SHIFT)
 | 
						|
#define TCR_IRGN0_NM_WBNWAC        (0x03 << TCR_IRGN0_SHIFT)
 | 
						|
 | 
						|
#define TCR_ORGN0_SHIFT            10
 | 
						|
#define TCR_ORGN0_NM_NC            (0x00 << TCR_ORGN0_SHIFT)
 | 
						|
#define TCR_ORGN0_NM_WBWAC         (0x01 << TCR_ORGN0_SHIFT)
 | 
						|
#define TCR_ORGN0_NM_WTC           (0x02 << TCR_ORGN0_SHIFT)
 | 
						|
#define TCR_ORGN0_NM_WBNWAC        (0x03 << TCR_ORGN0_SHIFT)
 | 
						|
 | 
						|
#define TCR_SH0_SHIFT              12
 | 
						|
#define TCR_SH0_NC                 (0x0 << TCR_SH0_SHIFT)
 | 
						|
#define TCR_SH0_OS                 (0x2 << TCR_SH0_SHIFT)
 | 
						|
#define TCR_SH0_IS                 (0x3 << TCR_SH0_SHIFT)
 | 
						|
 | 
						|
#define TCR_TG0_SHIFT              14
 | 
						|
#define TCR_TG0_4KB                (0x0 << TCR_TG0_SHIFT)
 | 
						|
#define TCR_TG0_64KB               (0x1 << TCR_TG0_SHIFT)
 | 
						|
#define TCR_TG0_16KB               (0x2 << TCR_TG0_SHIFT)
 | 
						|
 | 
						|
#define TCR_PS_SHIFT               16
 | 
						|
#define TCR_PS_4GB                 (0x0 << TCR_PS_SHIFT)
 | 
						|
#define TCR_PS_64GB                (0x1 << TCR_PS_SHIFT)
 | 
						|
#define TCR_PS_1TB                 (0x2 << TCR_PS_SHIFT)
 | 
						|
#define TCR_PS_4TB                 (0x3 << TCR_PS_SHIFT)
 | 
						|
#define TCR_PS_16TB                (0x4 << TCR_PS_SHIFT)
 | 
						|
#define TCR_PS_256TB               (0x5 << TCR_PS_SHIFT)
 | 
						|
 | 
						|
#define TCR_TBI_SHIFT              20
 | 
						|
#define TCR_TBI_USED               (0x0 << TCR_TBI_SHIFT)
 | 
						|
#define TCR_TBI_IGNORED            (0x1 << TCR_TBI_SHIFT)
 | 
						|
 | 
						|
#define DMA_DEFAULT_SIZE           (32 * MiB)
 | 
						|
#define TTB_DEFAULT_SIZE           0x100000
 | 
						|
 | 
						|
#define MB_SIZE			   (1UL << 20)
 | 
						|
 | 
						|
/* Initialize the MMU TTB tables using the mmu_ranges */
 | 
						|
uint64_t mmu_init(struct mmu_ranges *mmu_ranges);
 | 
						|
 | 
						|
/* Enable the mmu based on previous mmu_init(). */
 | 
						|
void mmu_enable(void);
 | 
						|
 | 
						|
/* Disable mmu */
 | 
						|
void mmu_disable(void);
 | 
						|
 | 
						|
/* Change a memory type for a range of bytes at runtime. */
 | 
						|
void mmu_config_range(void *start, size_t size, uint64_t tag);
 | 
						|
 | 
						|
/*
 | 
						|
 * Based on the memory ranges provided in coreboot tables,
 | 
						|
 * initialize the mmu_memranges used for mmu initialization
 | 
						|
 * cb_ranges -> Memory ranges present in cb tables
 | 
						|
 * mmu_ranges -> mmu_memranges initialized by this function
 | 
						|
 */
 | 
						|
struct mmu_memrange* mmu_init_ranges_from_sysinfo(struct memrange *cb_ranges,
 | 
						|
						  uint64_t ncb,
 | 
						|
						  struct mmu_ranges *mmu_ranges);
 | 
						|
 | 
						|
/*
 | 
						|
 * Functions for handling the initialization of memory ranges and enabling mmu
 | 
						|
 * before coreboot tables are parsed
 | 
						|
 */
 | 
						|
void mmu_presysinfo_memory_used(uint64_t base, uint64_t size);
 | 
						|
void mmu_presysinfo_enable(void);
 | 
						|
 | 
						|
/*
 | 
						|
 * Functions for exposing the used memory ranges to payloads. The ranges contain
 | 
						|
 * all used memory ranges that are actually used by payload. i.e. _start -> _end
 | 
						|
 * in linker script, the coreboot tables and framebuffer/DMA allocated in MMU
 | 
						|
 * initialization.
 | 
						|
 */
 | 
						|
const struct mmu_ranges *mmu_get_used_ranges(void);
 | 
						|
#endif // __ARCH_ARM64_MMU_H__
 |