This patch aims to make timestamps more consistent in naming, to follow one pattern. Until now there were many naming patterns: - TS_START_*/TS_END_* - TS_BEFORE_*/TS_AFTER_* - TS_*_START/TS_*_END This change also aims to indicate, that these timestamps can be used to create time-ranges, e.g. from TS_BOOTBLOCK_START to TS_BOOTBLOCK_END. Signed-off-by: Jakub Czapiga <jacz@semihalf.com> Change-Id: I533e32392224d9b67c37e6a67987b09bf1cf51c6 Reviewed-on: https://review.coreboot.org/c/coreboot/+/62019 Tested-by: build bot (Jenkins) <no-reply@coreboot.org> Reviewed-by: Tim Wawrzynczak <twawrzynczak@chromium.org> Reviewed-by: Yu-Ping Wu <yupingso@google.com> Reviewed-by: Raul Rangel <rrangel@chromium.org>
482 lines
13 KiB
C
482 lines
13 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
|
|
#include <console/console.h>
|
|
#include <commonlib/region.h>
|
|
#include <cf9_reset.h>
|
|
#include <string.h>
|
|
#include <arch/cpu.h>
|
|
#include <device/mmio.h>
|
|
#include <device/pci_ops.h>
|
|
#include <device/smbus_host.h>
|
|
#include <cbmem.h>
|
|
#include <timestamp.h>
|
|
#include <mrc_cache.h>
|
|
#include <southbridge/intel/bd82x6x/me.h>
|
|
#include <southbridge/intel/bd82x6x/pch.h>
|
|
#include <cpu/x86/msr.h>
|
|
#include <types.h>
|
|
|
|
#include "raminit_native.h"
|
|
#include "raminit_common.h"
|
|
#include "sandybridge.h"
|
|
|
|
/* FIXME: no support for 3-channel chipsets */
|
|
|
|
static void wait_txt_clear(void)
|
|
{
|
|
struct cpuid_result cp = cpuid_ext(1, 0);
|
|
|
|
/* Check if TXT is supported */
|
|
if (!(cp.ecx & (1 << 6)))
|
|
return;
|
|
|
|
/* Some TXT public bit */
|
|
if (!(read32((void *)0xfed30010) & 1))
|
|
return;
|
|
|
|
/* Wait for TXT clear */
|
|
while (!(read8((void *)0xfed40000) & (1 << 7)))
|
|
;
|
|
}
|
|
|
|
/* Disable a channel in ramctr_timing */
|
|
static void disable_channel(ramctr_timing *ctrl, int channel)
|
|
{
|
|
ctrl->rankmap[channel] = 0;
|
|
|
|
memset(&ctrl->rank_mirror[channel][0], 0, sizeof(ctrl->rank_mirror[0]));
|
|
|
|
ctrl->channel_size_mb[channel] = 0;
|
|
ctrl->cmd_stretch[channel] = 0;
|
|
ctrl->mad_dimm[channel] = 0;
|
|
memset(&ctrl->timings[channel][0], 0, sizeof(ctrl->timings[0]));
|
|
memset(&ctrl->info.dimm[channel][0], 0, sizeof(ctrl->info.dimm[0]));
|
|
}
|
|
|
|
static uint8_t nb_get_ecc_type(const uint32_t capid0_a)
|
|
{
|
|
return capid0_a & CAPID_ECCDIS ? MEMORY_ARRAY_ECC_NONE : MEMORY_ARRAY_ECC_SINGLE_BIT;
|
|
}
|
|
|
|
static uint16_t nb_slots_per_channel(const uint32_t capid0_a)
|
|
{
|
|
return !(capid0_a & CAPID_DDPCD) + 1;
|
|
}
|
|
|
|
static uint16_t nb_number_of_channels(const uint32_t capid0_a)
|
|
{
|
|
return !(capid0_a & CAPID_PDCD) + 1;
|
|
}
|
|
|
|
static uint32_t nb_max_chan_capacity_mib(const uint32_t capid0_a)
|
|
{
|
|
uint32_t ddrsz;
|
|
|
|
/* Values from documentation, which assume two DIMMs per channel */
|
|
switch (CAPID_DDRSZ(capid0_a)) {
|
|
case 1:
|
|
ddrsz = 8192;
|
|
break;
|
|
case 2:
|
|
ddrsz = 2048;
|
|
break;
|
|
case 3:
|
|
ddrsz = 512;
|
|
break;
|
|
default:
|
|
ddrsz = 16384;
|
|
break;
|
|
}
|
|
|
|
/* Account for the maximum number of DIMMs per channel */
|
|
return (ddrsz / 2) * nb_slots_per_channel(capid0_a);
|
|
}
|
|
|
|
/* Fill cbmem with information for SMBIOS type 16 and type 17 */
|
|
static void setup_sdram_meminfo(ramctr_timing *ctrl)
|
|
{
|
|
int channel, slot;
|
|
const u16 ddr_freq = (1000 << 8) / ctrl->tCK;
|
|
|
|
FOR_ALL_CHANNELS for (slot = 0; slot < NUM_SLOTS; slot++) {
|
|
enum cb_err ret = spd_add_smbios17(channel, slot, ddr_freq,
|
|
&ctrl->info.dimm[channel][slot]);
|
|
if (ret != CB_SUCCESS)
|
|
printk(BIOS_ERR, "RAMINIT: Failed to add SMBIOS17\n");
|
|
}
|
|
|
|
/* The 'spd_add_smbios17' function allocates this CBMEM area */
|
|
struct memory_info *m = cbmem_find(CBMEM_ID_MEMINFO);
|
|
if (m == NULL)
|
|
return;
|
|
|
|
const uint32_t capid0_a = pci_read_config32(HOST_BRIDGE, CAPID0_A);
|
|
|
|
const uint16_t channels = nb_number_of_channels(capid0_a);
|
|
|
|
m->ecc_type = nb_get_ecc_type(capid0_a);
|
|
m->max_capacity_mib = channels * nb_max_chan_capacity_mib(capid0_a);
|
|
m->number_of_devices = channels * nb_slots_per_channel(capid0_a);
|
|
}
|
|
|
|
/* Return CRC16 match for all SPDs */
|
|
static int verify_crc16_spds_ddr3(spd_raw_data *spd, ramctr_timing *ctrl)
|
|
{
|
|
int channel, slot, spd_slot;
|
|
int match = 1;
|
|
|
|
FOR_ALL_CHANNELS {
|
|
for (slot = 0; slot < NUM_SLOTS; slot++) {
|
|
spd_slot = 2 * channel + slot;
|
|
match &= ctrl->spd_crc[channel][slot] ==
|
|
spd_ddr3_calc_unique_crc(spd[spd_slot], sizeof(spd_raw_data));
|
|
}
|
|
}
|
|
return match;
|
|
}
|
|
|
|
void read_spd(spd_raw_data * spd, u8 addr, bool id_only)
|
|
{
|
|
int j;
|
|
if (id_only) {
|
|
for (j = 117; j < 128; j++)
|
|
(*spd)[j] = smbus_read_byte(addr, j);
|
|
} else {
|
|
for (j = 0; j < 256; j++)
|
|
(*spd)[j] = smbus_read_byte(addr, j);
|
|
}
|
|
}
|
|
|
|
static void dram_find_spds_ddr3(spd_raw_data *spd, ramctr_timing *ctrl)
|
|
{
|
|
int dimms = 0, ch_dimms;
|
|
int channel, slot, spd_slot;
|
|
bool can_use_ecc = ctrl->ecc_supported;
|
|
|
|
memset (ctrl->rankmap, 0, sizeof(ctrl->rankmap));
|
|
|
|
ctrl->extended_temperature_range = 1;
|
|
ctrl->auto_self_refresh = 1;
|
|
|
|
FOR_ALL_CHANNELS {
|
|
ctrl->channel_size_mb[channel] = 0;
|
|
|
|
ch_dimms = 0;
|
|
/* Count dimms on channel */
|
|
for (slot = 0; slot < NUM_SLOTS; slot++) {
|
|
spd_slot = 2 * channel + slot;
|
|
|
|
if (spd[spd_slot][SPD_MEMORY_TYPE] == SPD_MEMORY_TYPE_SDRAM_DDR3)
|
|
ch_dimms++;
|
|
}
|
|
|
|
for (slot = 0; slot < NUM_SLOTS; slot++) {
|
|
spd_slot = 2 * channel + slot;
|
|
printk(BIOS_DEBUG, "SPD probe channel%d, slot%d\n", channel, slot);
|
|
|
|
struct dimm_attr_ddr3_st *const dimm = &ctrl->info.dimm[channel][slot];
|
|
|
|
/* Search for XMP profile */
|
|
spd_xmp_decode_ddr3(dimm, spd[spd_slot],
|
|
DDR3_XMP_PROFILE_1);
|
|
|
|
if (dimm->dram_type != SPD_MEMORY_TYPE_SDRAM_DDR3) {
|
|
printram("No valid XMP profile found.\n");
|
|
spd_decode_ddr3(dimm, spd[spd_slot]);
|
|
|
|
} else if (ch_dimms > dimm->dimms_per_channel) {
|
|
printram(
|
|
"XMP profile supports %u DIMMs, but %u DIMMs are installed.\n",
|
|
dimm->dimms_per_channel, ch_dimms);
|
|
|
|
if (CONFIG(NATIVE_RAMINIT_IGNORE_XMP_MAX_DIMMS))
|
|
printk(BIOS_WARNING,
|
|
"XMP maximum DIMMs will be ignored.\n");
|
|
else
|
|
spd_decode_ddr3(dimm, spd[spd_slot]);
|
|
|
|
} else if (dimm->voltage != 1500) {
|
|
/* TODO: Support DDR3 voltages other than 1500mV */
|
|
printram("XMP profile's requested %u mV is unsupported.\n",
|
|
dimm->voltage);
|
|
|
|
if (CONFIG(NATIVE_RAMINIT_IGNORE_XMP_REQUESTED_VOLTAGE))
|
|
printk(BIOS_WARNING,
|
|
"XMP requested voltage will be ignored.\n");
|
|
else
|
|
spd_decode_ddr3(dimm, spd[spd_slot]);
|
|
}
|
|
|
|
/* Fill in CRC16 for MRC cache */
|
|
ctrl->spd_crc[channel][slot] =
|
|
spd_ddr3_calc_unique_crc(spd[spd_slot], sizeof(spd_raw_data));
|
|
|
|
if (dimm->dram_type != SPD_MEMORY_TYPE_SDRAM_DDR3) {
|
|
/* Mark DIMM as invalid */
|
|
dimm->ranks = 0;
|
|
dimm->size_mb = 0;
|
|
continue;
|
|
}
|
|
|
|
dram_print_spd_ddr3(dimm);
|
|
dimms++;
|
|
ctrl->rank_mirror[channel][slot * 2] = 0;
|
|
ctrl->rank_mirror[channel][slot * 2 + 1] = dimm->flags.pins_mirrored;
|
|
|
|
ctrl->channel_size_mb[channel] += dimm->size_mb;
|
|
|
|
if (!dimm->flags.is_ecc)
|
|
can_use_ecc = false;
|
|
|
|
ctrl->auto_self_refresh &= dimm->flags.asr;
|
|
|
|
ctrl->extended_temperature_range &= dimm->flags.ext_temp_refresh;
|
|
|
|
ctrl->rankmap[channel] |= ((1 << dimm->ranks) - 1) << (2 * slot);
|
|
|
|
printk(BIOS_DEBUG, "channel[%d] rankmap = 0x%x\n", channel,
|
|
ctrl->rankmap[channel]);
|
|
}
|
|
|
|
const u8 rc_0 = ctrl->info.dimm[channel][0].reference_card;
|
|
const u8 rc_1 = ctrl->info.dimm[channel][1].reference_card;
|
|
|
|
if (ch_dimms == NUM_SLOTS && rc_0 < 6 && rc_1 < 6) {
|
|
const int ref_card_offset_table[6][6] = {
|
|
{ 0, 0, 0, 0, 2, 2 },
|
|
{ 0, 0, 0, 0, 2, 2 },
|
|
{ 0, 0, 0, 0, 2, 2 },
|
|
{ 0, 0, 0, 0, 1, 1 },
|
|
{ 2, 2, 2, 1, 0, 0 },
|
|
{ 2, 2, 2, 1, 0, 0 },
|
|
};
|
|
ctrl->ref_card_offset[channel] = ref_card_offset_table[rc_0][rc_1];
|
|
} else {
|
|
ctrl->ref_card_offset[channel] = 0;
|
|
}
|
|
}
|
|
|
|
if (ctrl->ecc_forced || CONFIG(RAMINIT_ENABLE_ECC))
|
|
ctrl->ecc_enabled = can_use_ecc;
|
|
if (ctrl->ecc_forced && !ctrl->ecc_enabled)
|
|
die("ECC mode forced but non-ECC DIMM installed!");
|
|
printk(BIOS_DEBUG, "ECC is %s\n", ctrl->ecc_enabled ? "enabled" : "disabled");
|
|
|
|
ctrl->lanes = ctrl->ecc_enabled ? 9 : 8;
|
|
|
|
if (!dimms)
|
|
die("No DIMMs were found");
|
|
}
|
|
|
|
static void save_timings(ramctr_timing *ctrl)
|
|
{
|
|
/* Save the MRC S3 restore data to cbmem */
|
|
mrc_cache_stash_data(MRC_TRAINING_DATA, MRC_CACHE_VERSION, ctrl, sizeof(*ctrl));
|
|
}
|
|
|
|
static void reinit_ctrl(ramctr_timing *ctrl, const u32 cpuid)
|
|
{
|
|
/* Reset internal state */
|
|
memset(ctrl, 0, sizeof(*ctrl));
|
|
|
|
/* Get architecture */
|
|
ctrl->cpu = cpuid;
|
|
|
|
/* Get ECC support and mode */
|
|
ctrl->ecc_forced = get_host_ecc_forced();
|
|
ctrl->ecc_supported = ctrl->ecc_forced || get_host_ecc_cap();
|
|
printk(BIOS_DEBUG, "ECC supported: %s ECC forced: %s\n",
|
|
ctrl->ecc_supported ? "yes" : "no",
|
|
ctrl->ecc_forced ? "yes" : "no");
|
|
}
|
|
|
|
static void init_dram_ddr3(int s3resume, const u32 cpuid)
|
|
{
|
|
int me_uma_size, cbmem_was_inited, fast_boot, err;
|
|
ramctr_timing ctrl;
|
|
spd_raw_data spds[4];
|
|
size_t mrc_size;
|
|
ramctr_timing *ctrl_cached = NULL;
|
|
|
|
timestamp_add_now(TS_INITRAM_START);
|
|
|
|
mchbar_setbits32(SAPMCTL, 1 << 0);
|
|
|
|
/* Wait for ME to be ready */
|
|
intel_early_me_init();
|
|
me_uma_size = intel_early_me_uma_size();
|
|
|
|
printk(BIOS_DEBUG, "Starting native Platform init\n");
|
|
|
|
wait_txt_clear();
|
|
|
|
wrmsr(0x2e6, (msr_t) { .lo = 0, .hi = 0 });
|
|
|
|
const u32 sskpd = mchbar_read32(SSKPD); // !!! = 0x00000000
|
|
if ((pci_read_config16(SOUTHBRIDGE, 0xa2) & 0xa0) == 0x20 && sskpd && !s3resume) {
|
|
mchbar_write32(SSKPD, 0);
|
|
/* Need reset */
|
|
system_reset();
|
|
}
|
|
|
|
early_pch_init_native();
|
|
early_init_dmi();
|
|
early_thermal_init();
|
|
|
|
/* Try to find timings in MRC cache */
|
|
ctrl_cached = mrc_cache_current_mmap_leak(MRC_TRAINING_DATA,
|
|
MRC_CACHE_VERSION,
|
|
&mrc_size);
|
|
if (mrc_size < sizeof(ctrl))
|
|
ctrl_cached = NULL;
|
|
|
|
/* Before reusing training data, assert that the CPU has not been replaced */
|
|
if (ctrl_cached && cpuid != ctrl_cached->cpu) {
|
|
|
|
/* It is not really worrying on a cold boot, but fatal when resuming from S3 */
|
|
printk(s3resume ? BIOS_ALERT : BIOS_NOTICE,
|
|
"CPUID %x differs from stored CPUID %x, CPU was replaced!\n",
|
|
cpuid, ctrl_cached->cpu);
|
|
|
|
/* Invalidate the stored data, it likely does not apply to the current CPU */
|
|
ctrl_cached = NULL;
|
|
}
|
|
|
|
if (s3resume && !ctrl_cached) {
|
|
/* S3 resume is impossible, reset to come up cleanly */
|
|
system_reset();
|
|
}
|
|
|
|
/* Verify MRC cache for fast boot */
|
|
if (!s3resume && ctrl_cached) {
|
|
/* Load SPD unique information data. */
|
|
memset(spds, 0, sizeof(spds));
|
|
mainboard_get_spd(spds, 1);
|
|
|
|
/* check SPD CRC16 to make sure the DIMMs haven't been replaced */
|
|
fast_boot = verify_crc16_spds_ddr3(spds, ctrl_cached);
|
|
if (!fast_boot)
|
|
printk(BIOS_DEBUG, "Stored timings CRC16 mismatch.\n");
|
|
} else {
|
|
fast_boot = s3resume;
|
|
}
|
|
|
|
if (fast_boot) {
|
|
printk(BIOS_DEBUG, "Trying stored timings.\n");
|
|
memcpy(&ctrl, ctrl_cached, sizeof(ctrl));
|
|
|
|
err = try_init_dram_ddr3(&ctrl, fast_boot, s3resume, me_uma_size);
|
|
if (err) {
|
|
if (s3resume) {
|
|
/* Failed S3 resume, reset to come up cleanly */
|
|
system_reset();
|
|
}
|
|
/* No need to erase bad MRC cache here, it gets overwritten on a
|
|
successful boot */
|
|
printk(BIOS_ERR, "Stored timings are invalid !\n");
|
|
fast_boot = 0;
|
|
}
|
|
}
|
|
if (!fast_boot) {
|
|
/* Reset internal state */
|
|
reinit_ctrl(&ctrl, cpuid);
|
|
|
|
printk(BIOS_INFO, "ECC RAM %s.\n", ctrl.ecc_forced ? "required" :
|
|
ctrl.ecc_supported ? "supported" : "unsupported");
|
|
|
|
/* Get DDR3 SPD data */
|
|
memset(spds, 0, sizeof(spds));
|
|
mainboard_get_spd(spds, 0);
|
|
dram_find_spds_ddr3(spds, &ctrl);
|
|
|
|
err = try_init_dram_ddr3(&ctrl, fast_boot, s3resume, me_uma_size);
|
|
}
|
|
|
|
if (err) {
|
|
/* Fallback: disable failing channel */
|
|
printk(BIOS_ERR, "RAM training failed, trying fallback.\n");
|
|
printram("Disable failing channel.\n");
|
|
|
|
/* Reset internal state */
|
|
reinit_ctrl(&ctrl, cpuid);
|
|
|
|
/* Reset DDR3 frequency */
|
|
dram_find_spds_ddr3(spds, &ctrl);
|
|
|
|
/* Disable failing channel */
|
|
disable_channel(&ctrl, GET_ERR_CHANNEL(err));
|
|
|
|
err = try_init_dram_ddr3(&ctrl, fast_boot, s3resume, me_uma_size);
|
|
}
|
|
|
|
if (err)
|
|
die("raminit failed");
|
|
|
|
/* FIXME: should be hardware revision-dependent. The register only exists on IVB. */
|
|
mchbar_write32(CHANNEL_HASH, 0x00a030ce);
|
|
|
|
set_scrambling_seed(&ctrl);
|
|
|
|
if (!s3resume && ctrl.ecc_enabled)
|
|
channel_scrub(&ctrl);
|
|
|
|
set_normal_operation(&ctrl);
|
|
|
|
final_registers(&ctrl);
|
|
|
|
/* can't do this earlier because it needs to be done in normal operation */
|
|
if (CONFIG(DEBUG_RAM_SETUP) && !s3resume && ctrl.ecc_enabled) {
|
|
uint32_t i, tseg = pci_read_config32(HOST_BRIDGE, TSEGMB);
|
|
|
|
printk(BIOS_INFO, "RAMINIT: ECC scrub test on first channel up to 0x%x\n",
|
|
tseg);
|
|
|
|
/*
|
|
* This test helps to debug the ECC scrubbing.
|
|
* It likely tests every channel/rank, as rank interleave and enhanced
|
|
* interleave are enabled, but there's no guarantee for it.
|
|
*/
|
|
|
|
/* Skip first MB to avoid special case for A-seg and test up to TSEG */
|
|
for (i = 1; i < tseg >> 20; i++) {
|
|
for (int j = 0; j < 1 * MiB; j += 4096) {
|
|
uintptr_t addr = i * MiB + j;
|
|
if (read32((u32 *)addr) == 0)
|
|
continue;
|
|
|
|
printk(BIOS_ERR, "RAMINIT: ECC scrub: DRAM not cleared at"
|
|
" addr 0x%lx\n", addr);
|
|
break;
|
|
}
|
|
}
|
|
printk(BIOS_INFO, "RAMINIT: ECC scrub test done.\n");
|
|
}
|
|
|
|
/* Zone config */
|
|
dram_zones(&ctrl, 0);
|
|
|
|
intel_early_me_init_done(ME_INIT_STATUS_SUCCESS);
|
|
intel_early_me_status();
|
|
|
|
report_memory_config();
|
|
|
|
timestamp_add_now(TS_INITRAM_END);
|
|
|
|
cbmem_was_inited = !cbmem_recovery(s3resume);
|
|
if (!fast_boot)
|
|
save_timings(&ctrl);
|
|
if (s3resume && !cbmem_was_inited) {
|
|
/* Failed S3 resume, reset to come up cleanly */
|
|
system_reset();
|
|
}
|
|
|
|
if (!s3resume)
|
|
setup_sdram_meminfo(&ctrl);
|
|
}
|
|
|
|
void perform_raminit(int s3resume)
|
|
{
|
|
post_code(0x3a);
|
|
init_dram_ddr3(s3resume, cpu_get_cpuid());
|
|
}
|