/* SPDX-License-Identifier: GPL-2.0-only */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "raminit_native.h" #include "raminit_common.h" #include "sandybridge.h" /* FIXME: no support for 3-channel chipsets */ static void wait_txt_clear(void) { struct cpuid_result cp = cpuid_ext(1, 0); /* Check if TXT is supported */ if (!(cp.ecx & (1 << 6))) return; /* Some TXT public bit */ if (!(read32((void *)0xfed30010) & 1)) return; /* Wait for TXT clear */ while (!(read8((void *)0xfed40000) & (1 << 7))) ; } /* Disable a channel in ramctr_timing */ static void disable_channel(ramctr_timing *ctrl, int channel) { ctrl->rankmap[channel] = 0; memset(&ctrl->rank_mirror[channel][0], 0, sizeof(ctrl->rank_mirror[0])); ctrl->channel_size_mb[channel] = 0; ctrl->cmd_stretch[channel] = 0; ctrl->mad_dimm[channel] = 0; memset(&ctrl->timings[channel][0], 0, sizeof(ctrl->timings[0])); memset(&ctrl->info.dimm[channel][0], 0, sizeof(ctrl->info.dimm[0])); } static uint8_t nb_get_ecc_type(const uint32_t capid0_a) { return capid0_a & CAPID_ECCDIS ? MEMORY_ARRAY_ECC_NONE : MEMORY_ARRAY_ECC_SINGLE_BIT; } static uint16_t nb_slots_per_channel(const uint32_t capid0_a) { return !(capid0_a & CAPID_DDPCD) + 1; } static uint16_t nb_number_of_channels(const uint32_t capid0_a) { return !(capid0_a & CAPID_PDCD) + 1; } static uint32_t nb_max_chan_capacity_mib(const uint32_t capid0_a) { uint32_t ddrsz; /* Values from documentation, which assume two DIMMs per channel */ switch (CAPID_DDRSZ(capid0_a)) { case 1: ddrsz = 8192; break; case 2: ddrsz = 2048; break; case 3: ddrsz = 512; break; default: ddrsz = 16384; break; } /* Account for the maximum number of DIMMs per channel */ return (ddrsz / 2) * nb_slots_per_channel(capid0_a); } /* Fill cbmem with information for SMBIOS type 16 and type 17 */ static void setup_sdram_meminfo(ramctr_timing *ctrl) { int channel, slot; const u16 ddr_freq = (1000 << 8) / ctrl->tCK; FOR_ALL_CHANNELS for (slot = 0; slot < NUM_SLOTS; slot++) { enum cb_err ret = spd_add_smbios17(channel, slot, ddr_freq, &ctrl->info.dimm[channel][slot]); if (ret != CB_SUCCESS) printk(BIOS_ERR, "RAMINIT: Failed to add SMBIOS17\n"); } /* The 'spd_add_smbios17' function allocates this CBMEM area */ struct memory_info *m = cbmem_find(CBMEM_ID_MEMINFO); if (!m) return; const uint32_t capid0_a = pci_read_config32(HOST_BRIDGE, CAPID0_A); const uint16_t channels = nb_number_of_channels(capid0_a); m->ecc_type = nb_get_ecc_type(capid0_a); m->max_capacity_mib = channels * nb_max_chan_capacity_mib(capid0_a); m->number_of_devices = channels * nb_slots_per_channel(capid0_a); } /* Return CRC16 match for all SPDs */ static int verify_crc16_spds_ddr3(spd_raw_data *spd, ramctr_timing *ctrl) { int channel, slot, spd_slot; int match = 1; FOR_ALL_CHANNELS { for (slot = 0; slot < NUM_SLOTS; slot++) { spd_slot = 2 * channel + slot; match &= ctrl->spd_crc[channel][slot] == spd_ddr3_calc_unique_crc(spd[spd_slot], sizeof(spd_raw_data)); } } return match; } void read_spd(spd_raw_data * spd, u8 addr, bool id_only) { int j; if (id_only) { for (j = 117; j < 128; j++) (*spd)[j] = smbus_read_byte(addr, j); } else { for (j = 0; j < 256; j++) (*spd)[j] = smbus_read_byte(addr, j); } } static void dram_find_spds_ddr3(spd_raw_data *spd, ramctr_timing *ctrl) { int dimms = 0, ch_dimms; int channel, slot, spd_slot; bool can_use_ecc = ctrl->ecc_supported; memset (ctrl->rankmap, 0, sizeof(ctrl->rankmap)); ctrl->extended_temperature_range = 1; ctrl->auto_self_refresh = 1; FOR_ALL_CHANNELS { ctrl->channel_size_mb[channel] = 0; ch_dimms = 0; /* Count dimms on channel */ for (slot = 0; slot < NUM_SLOTS; slot++) { spd_slot = 2 * channel + slot; if (spd[spd_slot][SPD_MEMORY_TYPE] == SPD_MEMORY_TYPE_SDRAM_DDR3) ch_dimms++; } for (slot = 0; slot < NUM_SLOTS; slot++) { spd_slot = 2 * channel + slot; printk(BIOS_DEBUG, "SPD probe channel%d, slot%d\n", channel, slot); struct dimm_attr_ddr3_st *const dimm = &ctrl->info.dimm[channel][slot]; /* Search for XMP profile */ spd_xmp_decode_ddr3(dimm, spd[spd_slot], DDR3_XMP_PROFILE_1); if (dimm->dram_type != SPD_MEMORY_TYPE_SDRAM_DDR3) { printram("No valid XMP profile found.\n"); spd_decode_ddr3(dimm, spd[spd_slot]); } else if (ch_dimms > dimm->dimms_per_channel) { printram( "XMP profile supports %u DIMMs, but %u DIMMs are installed.\n", dimm->dimms_per_channel, ch_dimms); if (CONFIG(NATIVE_RAMINIT_IGNORE_XMP_MAX_DIMMS)) printk(BIOS_WARNING, "XMP maximum DIMMs will be ignored.\n"); else spd_decode_ddr3(dimm, spd[spd_slot]); } else if (dimm->voltage != 1500) { /* TODO: Support DDR3 voltages other than 1500mV */ printram("XMP profile's requested %u mV is unsupported.\n", dimm->voltage); if (CONFIG(NATIVE_RAMINIT_IGNORE_XMP_REQUESTED_VOLTAGE)) printk(BIOS_WARNING, "XMP requested voltage will be ignored.\n"); else spd_decode_ddr3(dimm, spd[spd_slot]); } /* Fill in CRC16 for MRC cache */ ctrl->spd_crc[channel][slot] = spd_ddr3_calc_unique_crc(spd[spd_slot], sizeof(spd_raw_data)); if (dimm->dram_type != SPD_MEMORY_TYPE_SDRAM_DDR3) { /* Mark DIMM as invalid */ dimm->ranks = 0; dimm->size_mb = 0; continue; } dram_print_spd_ddr3(dimm); dimms++; ctrl->rank_mirror[channel][slot * 2] = 0; ctrl->rank_mirror[channel][slot * 2 + 1] = dimm->flags.pins_mirrored; ctrl->channel_size_mb[channel] += dimm->size_mb; if (!dimm->flags.is_ecc) can_use_ecc = false; ctrl->auto_self_refresh &= dimm->flags.asr; ctrl->extended_temperature_range &= dimm->flags.ext_temp_refresh; ctrl->rankmap[channel] |= ((1 << dimm->ranks) - 1) << (2 * slot); printk(BIOS_DEBUG, "channel[%d] rankmap = 0x%x\n", channel, ctrl->rankmap[channel]); } const u8 rc_0 = ctrl->info.dimm[channel][0].reference_card; const u8 rc_1 = ctrl->info.dimm[channel][1].reference_card; if (ch_dimms == NUM_SLOTS && rc_0 < 6 && rc_1 < 6) { const int ref_card_offset_table[6][6] = { { 0, 0, 0, 0, 2, 2 }, { 0, 0, 0, 0, 2, 2 }, { 0, 0, 0, 0, 2, 2 }, { 0, 0, 0, 0, 1, 1 }, { 2, 2, 2, 1, 0, 0 }, { 2, 2, 2, 1, 0, 0 }, }; ctrl->ref_card_offset[channel] = ref_card_offset_table[rc_0][rc_1]; } else { ctrl->ref_card_offset[channel] = 0; } } if (ctrl->ecc_forced || CONFIG(RAMINIT_ENABLE_ECC)) ctrl->ecc_enabled = can_use_ecc; if (ctrl->ecc_forced && !ctrl->ecc_enabled) die("ECC mode forced but non-ECC DIMM installed!"); printk(BIOS_DEBUG, "ECC is %s\n", ctrl->ecc_enabled ? "enabled" : "disabled"); ctrl->lanes = ctrl->ecc_enabled ? 9 : 8; if (!dimms) die("No DIMMs were found"); } static void save_timings(ramctr_timing *ctrl) { /* Save the MRC S3 restore data to cbmem */ mrc_cache_stash_data(MRC_TRAINING_DATA, MRC_CACHE_VERSION, ctrl, sizeof(*ctrl)); } static void reinit_ctrl(ramctr_timing *ctrl, const u32 cpuid) { /* Reset internal state */ memset(ctrl, 0, sizeof(*ctrl)); /* Get architecture */ ctrl->cpu = cpuid; /* Get ECC support and mode */ ctrl->ecc_forced = get_host_ecc_forced(); ctrl->ecc_supported = ctrl->ecc_forced || get_host_ecc_cap(); printk(BIOS_DEBUG, "ECC supported: %s ECC forced: %s\n", ctrl->ecc_supported ? "yes" : "no", ctrl->ecc_forced ? "yes" : "no"); } static void init_dram_ddr3(int s3resume, const u32 cpuid) { int me_uma_size, cbmem_was_inited, fast_boot, err; ramctr_timing ctrl; spd_raw_data spds[4]; size_t mrc_size; ramctr_timing *ctrl_cached = NULL; timestamp_add_now(TS_INITRAM_START); mchbar_setbits32(SAPMCTL, 1 << 0); /* Wait for ME to be ready */ intel_early_me_init(); me_uma_size = intel_early_me_uma_size(); printk(BIOS_DEBUG, "Starting native Platform init\n"); wait_txt_clear(); wrmsr(0x2e6, (msr_t) { .lo = 0, .hi = 0 }); const u32 sskpd = mchbar_read32(SSKPD); // !!! = 0x00000000 if ((pci_read_config16(SOUTHBRIDGE, 0xa2) & 0xa0) == 0x20 && sskpd && !s3resume) { mchbar_write32(SSKPD, 0); /* Need reset */ system_reset(); } early_pch_init_native(); early_init_dmi(); early_thermal_init(); /* Try to find timings in MRC cache */ ctrl_cached = mrc_cache_current_mmap_leak(MRC_TRAINING_DATA, MRC_CACHE_VERSION, &mrc_size); if (mrc_size < sizeof(ctrl)) ctrl_cached = NULL; /* Before reusing training data, assert that the CPU has not been replaced */ if (ctrl_cached && cpuid != ctrl_cached->cpu) { /* It is not really worrying on a cold boot, but fatal when resuming from S3 */ printk(s3resume ? BIOS_ALERT : BIOS_NOTICE, "CPUID %x differs from stored CPUID %x, CPU was replaced!\n", cpuid, ctrl_cached->cpu); /* Invalidate the stored data, it likely does not apply to the current CPU */ ctrl_cached = NULL; } if (s3resume && !ctrl_cached) { /* S3 resume is impossible, reset to come up cleanly */ system_reset(); } /* Verify MRC cache for fast boot */ if (!s3resume && ctrl_cached) { /* Load SPD unique information data. */ memset(spds, 0, sizeof(spds)); mainboard_get_spd(spds, 1); /* check SPD CRC16 to make sure the DIMMs haven't been replaced */ fast_boot = verify_crc16_spds_ddr3(spds, ctrl_cached); if (!fast_boot) printk(BIOS_DEBUG, "Stored timings CRC16 mismatch.\n"); } else { fast_boot = s3resume; } if (fast_boot) { printk(BIOS_DEBUG, "Trying stored timings.\n"); memcpy(&ctrl, ctrl_cached, sizeof(ctrl)); err = try_init_dram_ddr3(&ctrl, fast_boot, s3resume, me_uma_size); if (err) { if (s3resume) { /* Failed S3 resume, reset to come up cleanly */ system_reset(); } /* No need to erase bad MRC cache here, it gets overwritten on a successful boot */ printk(BIOS_ERR, "Stored timings are invalid !\n"); fast_boot = 0; } } if (!fast_boot) { /* Reset internal state */ reinit_ctrl(&ctrl, cpuid); printk(BIOS_INFO, "ECC RAM %s.\n", ctrl.ecc_forced ? "required" : ctrl.ecc_supported ? "supported" : "unsupported"); /* Get DDR3 SPD data */ memset(spds, 0, sizeof(spds)); mainboard_get_spd(spds, 0); dram_find_spds_ddr3(spds, &ctrl); err = try_init_dram_ddr3(&ctrl, fast_boot, s3resume, me_uma_size); } if (err) { /* Fallback: disable failing channel */ printk(BIOS_ERR, "RAM training failed, trying fallback.\n"); printram("Disable failing channel.\n"); /* Reset internal state */ reinit_ctrl(&ctrl, cpuid); /* Reset DDR3 frequency */ dram_find_spds_ddr3(spds, &ctrl); /* Disable failing channel */ disable_channel(&ctrl, GET_ERR_CHANNEL(err)); err = try_init_dram_ddr3(&ctrl, fast_boot, s3resume, me_uma_size); } if (err) die("raminit failed"); /* FIXME: should be hardware revision-dependent. The register only exists on IVB. */ mchbar_write32(CHANNEL_HASH, 0x00a030ce); set_scrambling_seed(&ctrl); if (!s3resume && ctrl.ecc_enabled) channel_scrub(&ctrl); set_normal_operation(&ctrl); final_registers(&ctrl); /* can't do this earlier because it needs to be done in normal operation */ if (CONFIG(DEBUG_RAM_SETUP) && !s3resume && ctrl.ecc_enabled) { uint32_t i, tseg = pci_read_config32(HOST_BRIDGE, TSEGMB); printk(BIOS_INFO, "RAMINIT: ECC scrub test on first channel up to 0x%x\n", tseg); /* * This test helps to debug the ECC scrubbing. * It likely tests every channel/rank, as rank interleave and enhanced * interleave are enabled, but there's no guarantee for it. */ /* Skip first MB to avoid special case for A-seg and test up to TSEG */ for (i = 1; i < tseg >> 20; i++) { for (int j = 0; j < 1 * MiB; j += 4096) { uintptr_t addr = i * MiB + j; if (read32((u32 *)addr) == 0) continue; printk(BIOS_ERR, "RAMINIT: ECC scrub: DRAM not cleared at" " addr 0x%lx\n", addr); break; } } printk(BIOS_INFO, "RAMINIT: ECC scrub test done.\n"); } /* Zone config */ dram_zones(&ctrl, 0); intel_early_me_init_done(ME_INIT_STATUS_SUCCESS); intel_early_me_status(); report_memory_config(); timestamp_add_now(TS_INITRAM_END); cbmem_was_inited = !cbmem_recovery(s3resume); if (!fast_boot) save_timings(&ctrl); if (s3resume && !cbmem_was_inited) { /* Failed S3 resume, reset to come up cleanly */ system_reset(); } if (!s3resume) setup_sdram_meminfo(&ctrl); } void perform_raminit(int s3resume) { post_code(0x3a); init_dram_ddr3(s3resume, cpu_get_cpuid()); }