This adds inline wrappers to read the L2 cache auxiliary control register (L2ACTLR). Signed-off-by: David Hendricks <dhendrix@chromium.org> Change-Id: Iec603d7c738426232f7ce3a4a474d01c85fa3f2f Reviewed-on: https://gerrit.chromium.org/gerrit/64861 Commit-Queue: David Hendricks <dhendrix@chromium.org> Reviewed-by: David Hendricks <dhendrix@chromium.org> Tested-by: David Hendricks <dhendrix@chromium.org> Reviewed-on: http://review.coreboot.org/4437 Tested-by: build bot (Jenkins) Reviewed-by: Patrick Georgi <patrick@georgi-clan.de>
		
			
				
	
	
		
			325 lines
		
	
	
		
			9.0 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			325 lines
		
	
	
		
			9.0 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * This file is part of the coreboot project.
 | 
						|
 *
 | 
						|
 * Copyright 2013 Google Inc.
 | 
						|
 *
 | 
						|
 * Redistribution and use in source and binary forms, with or without
 | 
						|
 * modification, are permitted provided that the following conditions
 | 
						|
 * are met:
 | 
						|
 * 1. Redistributions of source code must retain the above copyright
 | 
						|
 *    notice, this list of conditions and the following disclaimer.
 | 
						|
 * 2. Redistributions in binary form must reproduce the above copyright
 | 
						|
 *    notice, this list of conditions and the following disclaimer in the
 | 
						|
 *    documentation and/or other materials provided with the distribution.
 | 
						|
 * 3. The name of the author may not be used to endorse or promote products
 | 
						|
 *    derived from this software without specific prior written permission.
 | 
						|
 *
 | 
						|
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
 | 
						|
 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 | 
						|
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 | 
						|
 * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
 | 
						|
 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 | 
						|
 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
 | 
						|
 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 | 
						|
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 | 
						|
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 | 
						|
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 | 
						|
 * SUCH DAMAGE.
 | 
						|
 *
 | 
						|
 * cache.h: Cache maintenance API for ARMv7
 | 
						|
 */
 | 
						|
 | 
						|
#ifndef ARMV7_CACHE_H
 | 
						|
#define ARMV7_CACHE_H
 | 
						|
 | 
						|
#include <stdint.h>
 | 
						|
 | 
						|
/* SCTLR bits */
 | 
						|
#define SCTLR_M		(1 << 0)	/* MMU enable			*/
 | 
						|
#define SCTLR_A		(1 << 1)	/* Alignment check enable	*/
 | 
						|
#define SCTLR_C		(1 << 2)	/* Data/unified cache enable	*/
 | 
						|
/* Bits 4:3 are reserved */
 | 
						|
#define SCTLR_CP15BEN	(1 << 5)	/* CP15 barrier enable		*/
 | 
						|
/* Bit 6 is reserved */
 | 
						|
#define SCTLR_B		(1 << 7)	/* Endianness			*/
 | 
						|
/* Bits 9:8 */
 | 
						|
#define SCTLR_SW	(1 << 10)	/* SWP and SWPB enable		*/
 | 
						|
#define SCTLR_Z		(1 << 11)	/* Branch prediction enable	*/
 | 
						|
#define SCTLR_I		(1 << 12)	/* Instruction cache enable	*/
 | 
						|
#define SCTLR_V		(1 << 13)	/* Low/high exception vectors 	*/
 | 
						|
#define SCTLR_RR  	(1 << 14)	/* Round Robin select		*/
 | 
						|
/* Bits 16:15 are reserved */
 | 
						|
#define SCTLR_HA	(1 << 17)	/* Hardware Access flag enable	*/
 | 
						|
/* Bit 18 is reserved */
 | 
						|
/* Bits 20:19 reserved virtualization not supported */
 | 
						|
#define SCTLR_WXN	(1 << 19)	/* Write permission implies XN	*/
 | 
						|
#define SCTLR_UWXN	(1 << 20)	/* Unprivileged write permission
 | 
						|
					   implies PL1 XN		*/
 | 
						|
#define SCTLR_FI	(1 << 21)	/* Fast interrupt config enable	*/
 | 
						|
#define SCTLR_U		(1 << 22)	/* Unaligned access behavior	*/
 | 
						|
#define SCTLR_VE	(1 << 24)	/* Interrupt vectors enable	*/
 | 
						|
#define SCTLR_EE	(1 << 25)	/* Exception endianness		*/
 | 
						|
/* Bit 26 is reserved */
 | 
						|
#define SCTLR_NMFI	(1 << 27)	/* Non-maskable FIQ support	*/
 | 
						|
#define SCTLR_TRE	(1 << 28)	/* TEX remap enable		*/
 | 
						|
#define SCTLR_AFE	(1 << 29)	/* Access flag enable		*/
 | 
						|
#define SCTLR_TE	(1 << 30)	/* Thumb exception enable	*/
 | 
						|
/* Bit 31 is reserved */
 | 
						|
 | 
						|
/*
 | 
						|
 * Sync primitives
 | 
						|
 */
 | 
						|
 | 
						|
/* data memory barrier */
 | 
						|
static inline void dmb(void)
 | 
						|
{
 | 
						|
	asm volatile ("dmb" : : : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* data sync barrier */
 | 
						|
static inline void dsb(void)
 | 
						|
{
 | 
						|
	asm volatile ("dsb" : : : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* instruction sync barrier */
 | 
						|
static inline void isb(void)
 | 
						|
{
 | 
						|
	asm volatile ("isb" : : : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Low-level TLB maintenance operations
 | 
						|
 */
 | 
						|
 | 
						|
/* invalidate entire data TLB */
 | 
						|
static inline void dtlbiall(void)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c8, c6, 0" : : "r" (0) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* invalidate entire instruction TLB */
 | 
						|
static inline void itlbiall(void)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c8, c5, 0" : : "r" (0));
 | 
						|
}
 | 
						|
 | 
						|
/* invalidate entire unified TLB */
 | 
						|
static inline void tlbiall(void)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c8, c7, 0" : : "r" (0) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* write data access control register (DACR) */
 | 
						|
static inline void write_dacr(uint32_t val)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c3, c0, 0" : : "r" (val));
 | 
						|
}
 | 
						|
 | 
						|
/* write translation table base register 0 (TTBR0) */
 | 
						|
static inline void write_ttbr0(uint32_t val)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c2, c0, 0" : : "r" (val) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* read translation table base control register (TTBCR) */
 | 
						|
static inline uint32_t read_ttbcr(void)
 | 
						|
{
 | 
						|
	uint32_t val = 0;
 | 
						|
	asm volatile ("mrc p15, 0, %0, c2, c0, 2" : "=r" (val));
 | 
						|
	return val;
 | 
						|
}
 | 
						|
 | 
						|
/* write translation table base control register (TTBCR) */
 | 
						|
static inline void write_ttbcr(uint32_t val)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c2, c0, 2" : : "r" (val) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Low-level cache maintenance operations
 | 
						|
 */
 | 
						|
 | 
						|
/* branch predictor invalidate all */
 | 
						|
static inline void bpiall(void)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c7, c5, 6" : : "r" (0));
 | 
						|
}
 | 
						|
 | 
						|
/* data cache clean and invalidate by MVA to PoC */
 | 
						|
static inline void dccimvac(unsigned long mva)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c7, c14, 1" : : "r" (mva) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* data cache invalidate by set/way */
 | 
						|
static inline void dccisw(uint32_t val)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c7, c14, 2" : : "r" (val) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* data cache clean by MVA to PoC */
 | 
						|
static inline void dccmvac(unsigned long mva)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c7, c10, 1" : : "r" (mva) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* data cache invalidate by MVA to PoC */
 | 
						|
static inline void dcimvac(unsigned long mva)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c7, c6, 1" : : "r" (mva) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* data cache invalidate by set/way */
 | 
						|
static inline void dcisw(uint32_t val)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c7, c6, 2" : : "r" (val) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
/* instruction cache invalidate all by PoU */
 | 
						|
static inline void iciallu(void)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c7, c5, 0" : : "r" (0));
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Cache co-processor (CP15) access functions
 | 
						|
 */
 | 
						|
 | 
						|
/* read cache level ID register (CLIDR) */
 | 
						|
static inline uint32_t read_clidr(void)
 | 
						|
{
 | 
						|
	uint32_t val = 0;
 | 
						|
	asm volatile ("mrc p15, 1, %0, c0, c0, 1" : "=r" (val));
 | 
						|
	return val;
 | 
						|
}
 | 
						|
 | 
						|
/* read cache size ID register register (CCSIDR) */
 | 
						|
static inline uint32_t read_ccsidr(void)
 | 
						|
{
 | 
						|
	uint32_t val = 0;
 | 
						|
	asm volatile ("mrc p15, 1, %0, c0, c0, 0" : "=r" (val));
 | 
						|
	return val;
 | 
						|
}
 | 
						|
 | 
						|
/* read cache size selection register (CSSELR) */
 | 
						|
static inline uint32_t read_csselr(void)
 | 
						|
{
 | 
						|
	uint32_t val = 0;
 | 
						|
	asm volatile ("mrc p15, 2, %0, c0, c0, 0" : "=r" (val));
 | 
						|
	return val;
 | 
						|
}
 | 
						|
 | 
						|
/* write to cache size selection register (CSSELR) */
 | 
						|
static inline void write_csselr(uint32_t val)
 | 
						|
{
 | 
						|
	/*
 | 
						|
	 * Bits [3:1] - Cache level + 1 (0b000 = L1, 0b110 = L7, 0b111 is rsvd)
 | 
						|
	 * Bit 0 - 0 = data or unified cache, 1 = instruction cache
 | 
						|
	 */
 | 
						|
	asm volatile ("mcr p15, 2, %0, c0, c0, 0" : : "r" (val));
 | 
						|
	isb();	/* ISB to sync the change to CCSIDR */
 | 
						|
}
 | 
						|
 | 
						|
/* read L2 control register (L2CTLR) */
 | 
						|
static inline uint32_t read_l2ctlr(void)
 | 
						|
{
 | 
						|
	uint32_t val = 0;
 | 
						|
	asm volatile ("mrc p15, 1, %0, c9, c0, 2" : "=r" (val));
 | 
						|
	return val;
 | 
						|
}
 | 
						|
 | 
						|
/* write L2 control register (L2CTLR) */
 | 
						|
static inline void write_l2ctlr(uint32_t val)
 | 
						|
{
 | 
						|
	/*
 | 
						|
	 * Note: L2CTLR can only be written when the L2 memory system
 | 
						|
	 * is idle, ie before the MMU is enabled.
 | 
						|
	 */
 | 
						|
	asm volatile("mcr p15, 1, %0, c9, c0, 2" : : "r" (val) : "memory" );
 | 
						|
	isb();
 | 
						|
}
 | 
						|
 | 
						|
/* read L2 Auxiliary Control Register (L2ACTLR) */
 | 
						|
static inline uint32_t read_l2actlr(void)
 | 
						|
{
 | 
						|
	uint32_t val = 0;
 | 
						|
	asm volatile ("mrc p15, 1, %0, c15, c0, 0" : "=r" (val));
 | 
						|
	return val;
 | 
						|
}
 | 
						|
 | 
						|
/* write L2 Auxiliary Control Register (L2ACTLR) */
 | 
						|
static inline void write_l2actlr(uint32_t val)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 1, %0, c15, c0, 0" : : "r" (val) : "memory" );
 | 
						|
	isb();
 | 
						|
}
 | 
						|
 | 
						|
/* read system control register (SCTLR) */
 | 
						|
static inline uint32_t read_sctlr(void)
 | 
						|
{
 | 
						|
	uint32_t val;
 | 
						|
	asm volatile ("mrc p15, 0, %0, c1, c0, 0" : "=r" (val));
 | 
						|
	return val;
 | 
						|
}
 | 
						|
 | 
						|
/* write system control register (SCTLR) */
 | 
						|
static inline void write_sctlr(uint32_t val)
 | 
						|
{
 | 
						|
	asm volatile ("mcr p15, 0, %0, c1, c0, 0" : : "r" (val) : "cc");
 | 
						|
	isb();
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Cache maintenance API
 | 
						|
 */
 | 
						|
 | 
						|
/* dcache clean and invalidate all (on current level given by CCSELR) */
 | 
						|
void dcache_clean_invalidate_all(void);
 | 
						|
 | 
						|
/* dcache clean by modified virtual address to PoC */
 | 
						|
void dcache_clean_by_mva(unsigned long addr, unsigned long len);
 | 
						|
 | 
						|
/* dcache clean and invalidate by modified virtual address to PoC */
 | 
						|
void dcache_clean_invalidate_by_mva(unsigned long addr, unsigned long len);
 | 
						|
 | 
						|
/* dcache invalidate by modified virtual address to PoC */
 | 
						|
void dcache_invalidate_by_mva(unsigned long addr, unsigned long len);
 | 
						|
 | 
						|
/* dcache invalidate all (on current level given by CCSELR) */
 | 
						|
void dcache_invalidate_all(void);
 | 
						|
 | 
						|
/* dcache and MMU disable */
 | 
						|
void dcache_mmu_disable(void);
 | 
						|
 | 
						|
/* dcache and MMU enable */
 | 
						|
void dcache_mmu_enable(void);
 | 
						|
 | 
						|
/* icache invalidate all (on current level given by CSSELR) */
 | 
						|
void icache_invalidate_all(void);
 | 
						|
 | 
						|
/* tlb invalidate all */
 | 
						|
void tlb_invalidate_all(void);
 | 
						|
 | 
						|
/*
 | 
						|
 * Generalized setup/init functions
 | 
						|
 */
 | 
						|
 | 
						|
/* invalidate all caches on ARMv7 */
 | 
						|
void armv7_invalidate_caches(void);
 | 
						|
 | 
						|
/* mmu initialization (set page table address, set permissions, etc) */
 | 
						|
void mmu_init(void);
 | 
						|
 | 
						|
enum dcache_policy {
 | 
						|
	DCACHE_OFF,
 | 
						|
	DCACHE_WRITEBACK,
 | 
						|
	DCACHE_WRITETHROUGH,
 | 
						|
};
 | 
						|
 | 
						|
/* mmu range configuration (set dcache policy) */
 | 
						|
void mmu_config_range(unsigned long start_mb, unsigned long size_mb,
 | 
						|
						enum dcache_policy policy);
 | 
						|
 | 
						|
#endif /* ARMV7_CACHE_H */
 |