More WIP on new memory map
This commit is contained in:
parent
1f211b37df
commit
d0aa4c722d
@ -47,7 +47,6 @@ env:
|
|||||||
- CONFIG=DualCoreConfig
|
- CONFIG=DualCoreConfig
|
||||||
- CONFIG=MemtestConfig
|
- CONFIG=MemtestConfig
|
||||||
- CONFIG=FancyMemtestConfig
|
- CONFIG=FancyMemtestConfig
|
||||||
- CONFIG=MemoryMuxMemtestConfig
|
|
||||||
- CONFIG=BroadcastRegressionTestConfig
|
- CONFIG=BroadcastRegressionTestConfig
|
||||||
- CONFIG=CacheRegressionTestConfig
|
- CONFIG=CacheRegressionTestConfig
|
||||||
- CONFIG=NastiConverterTestConfig
|
- CONFIG=NastiConverterTestConfig
|
||||||
|
@ -34,7 +34,6 @@ class htif_emulator_t : public htif_pthread_t
|
|||||||
void start()
|
void start()
|
||||||
{
|
{
|
||||||
set_clock_divisor(5, 2);
|
set_clock_divisor(5, 2);
|
||||||
write_cr(-1, UNCORE_SCR__MEMORY_CHANNEL_MUX_SELECT__OFFSET, memory_channel_mux_select);
|
|
||||||
htif_pthread_t::start();
|
htif_pthread_t::start();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -14,18 +14,6 @@ extern "C" {
|
|||||||
|
|
||||||
extern int vcs_main(int argc, char** argv);
|
extern int vcs_main(int argc, char** argv);
|
||||||
|
|
||||||
static const int MEMORY_CHANNEL_MUX_CONFIGS[] = {
|
|
||||||
#ifdef MEMORY_CHANNEL_MUX_CONFIGS__0
|
|
||||||
MEMORY_CHANNEL_MUX_CONFIGS__0,
|
|
||||||
#endif
|
|
||||||
#ifdef MEMORY_CHANNEL_MUX_CONFIGS__1
|
|
||||||
MEMORY_CHANNEL_MUX_CONFIGS__1,
|
|
||||||
#endif
|
|
||||||
#ifdef MEMORY_CHANNEL_MUX_CONFIGS__2
|
|
||||||
#error "Add a preprocessor repeat macro"
|
|
||||||
#endif
|
|
||||||
};
|
|
||||||
|
|
||||||
static htif_emulator_t* htif;
|
static htif_emulator_t* htif;
|
||||||
static unsigned htif_bytes = HTIF_WIDTH / 8;
|
static unsigned htif_bytes = HTIF_WIDTH / 8;
|
||||||
static mm_t* mm[N_MEM_CHANNELS];
|
static mm_t* mm[N_MEM_CHANNELS];
|
||||||
@ -54,21 +42,19 @@ int main(int argc, char** argv)
|
|||||||
memory_channel_mux_select = atoi(argv[i]+27);
|
memory_channel_mux_select = atoi(argv[i]+27);
|
||||||
}
|
}
|
||||||
|
|
||||||
int enabled_mem_channels = MEMORY_CHANNEL_MUX_CONFIGS[memory_channel_mux_select];
|
|
||||||
|
|
||||||
htif = new htif_emulator_t(memsz_mb,
|
htif = new htif_emulator_t(memsz_mb,
|
||||||
std::vector<std::string>(argv + 1, argv + argc));
|
std::vector<std::string>(argv + 1, argv + argc));
|
||||||
|
|
||||||
for (int i=0; i<N_MEM_CHANNELS; i++) {
|
for (int i=0; i<N_MEM_CHANNELS; i++) {
|
||||||
mm[i] = dramsim ? (mm_t*)(new mm_dramsim2_t) : (mm_t*)(new mm_magic_t);
|
mm[i] = dramsim ? (mm_t*)(new mm_dramsim2_t) : (mm_t*)(new mm_magic_t);
|
||||||
mm[i]->init(MEM_SIZE / enabled_mem_channels, MEM_DATA_BITS / 8, CACHE_BLOCK_BYTES);
|
mm[i]->init(MEM_SIZE / N_MEM_CHANNELS, MEM_DATA_BITS / 8, CACHE_BLOCK_BYTES);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (loadmem) {
|
if (loadmem) {
|
||||||
void *mems[N_MEM_CHANNELS];
|
void *mems[N_MEM_CHANNELS];
|
||||||
for (int i = 0; i < N_MEM_CHANNELS; i++)
|
for (int i = 0; i < N_MEM_CHANNELS; i++)
|
||||||
mems[i] = mm[i]->get_data();
|
mems[i] = mm[i]->get_data();
|
||||||
load_mem(mems, loadmem, CACHE_BLOCK_BYTES, enabled_mem_channels);
|
load_mem(mems, loadmem, CACHE_BLOCK_BYTES, N_MEM_CHANNELS);
|
||||||
}
|
}
|
||||||
|
|
||||||
vcs_main(argc, argv);
|
vcs_main(argc, argv);
|
||||||
|
@ -1 +1 @@
|
|||||||
Subproject commit d9c4e92a21de34515f745893bcb4e76ed8f1cd60
|
Subproject commit 30ee6af52cada4fd8579f0654983ac73c1869473
|
2
rocket
2
rocket
@ -1 +1 @@
|
|||||||
Subproject commit 43baccd10a23a4e1e11b3ee0c571bae1729ae035
|
Subproject commit 4aca3088510183e344994ce0fc4206e95083a5a7
|
@ -24,36 +24,48 @@ class DefaultConfig extends Config (
|
|||||||
topDefinitions = { (pname,site,here) =>
|
topDefinitions = { (pname,site,here) =>
|
||||||
type PF = PartialFunction[Any,Any]
|
type PF = PartialFunction[Any,Any]
|
||||||
def findBy(sname:Any):Any = here[PF](site[Any](sname))(pname)
|
def findBy(sname:Any):Any = here[PF](site[Any](sname))(pname)
|
||||||
def genCsrAddrMap: AddrMap = {
|
lazy val internalIOAddrMap: AddrMap = {
|
||||||
val deviceTree = AddrMapEntry("devicetree", MemSize(1 << 15, AddrMapConsts.R))
|
val deviceTree = AddrMapEntry("configstring", MemSize(1<<13, 1<<12, MemAttr(AddrMapProt.R)))
|
||||||
val rtc = AddrMapEntry("rtc", MemSize(1 << 12, AddrMapConsts.RW))
|
val rtc = AddrMapEntry("rtc", MemSize(1<<12, 1<<12, MemAttr(AddrMapProt.RW)))
|
||||||
new AddrMap(Seq(deviceTree, rtc))
|
new AddrMap(Seq(deviceTree, rtc))
|
||||||
}
|
}
|
||||||
|
lazy val globalAddrMap: AddrMap = {
|
||||||
|
val memSize = 1L << 30
|
||||||
|
val memAlign = 1L << 31
|
||||||
|
val extIOSize = 1L << 29
|
||||||
|
val mem = MemSize(memSize, memAlign, MemAttr(AddrMapProt.RWX, true))
|
||||||
|
val io = AddrMap(
|
||||||
|
AddrMapEntry("int", MemSubmap(internalIOAddrMap.computeSize, internalIOAddrMap)),
|
||||||
|
AddrMapEntry("ext", MemSize(extIOSize, extIOSize, MemAttr(AddrMapProt.RWX))))
|
||||||
|
Dump("MEM_SIZE", memSize)
|
||||||
|
AddrMap(
|
||||||
|
AddrMapEntry("mem", mem),
|
||||||
|
AddrMapEntry("io", MemSubmap(io.computeSize, io)))
|
||||||
|
}
|
||||||
def makeConfigString() = {
|
def makeConfigString() = {
|
||||||
val addrMap = new AddrHashMap(site(GlobalAddrMap))
|
val addrMap = new AddrHashMap(globalAddrMap)
|
||||||
val xLen = site(XLen)
|
val xLen = site(XLen)
|
||||||
val res = new StringBuilder
|
val res = new StringBuilder
|
||||||
val memSize = addrMap("mem").size
|
|
||||||
val rtcAddr = addrMap("conf:rtc").start
|
|
||||||
res append "platform {\n"
|
res append "platform {\n"
|
||||||
res append " vendor ucb;\n"
|
res append " vendor ucb;\n"
|
||||||
res append " arch rocket;\n"
|
res append " arch rocket;\n"
|
||||||
res append "};\n"
|
res append "};\n"
|
||||||
res append "rtc {\n"
|
res append "rtc {\n"
|
||||||
res append s" addr 0x${rtcAddr.toString(16)};\n"
|
res append s" addr 0x${addrMap("io:int:rtc").start.toString(16)};\n"
|
||||||
res append "};\n"
|
res append "};\n"
|
||||||
res append "ram {\n"
|
res append "ram {\n"
|
||||||
res append " 0 {\n"
|
res append " 0 {\n"
|
||||||
res append " addr 0;\n"
|
res append s" addr 0x${addrMap("mem").start.toString(16)};\n"
|
||||||
res append s" size 0x${memSize.toString(16)};\n"
|
res append s" size 0x${addrMap("mem").region.size.toString(16)};\n"
|
||||||
res append " };\n"
|
res append " };\n"
|
||||||
res append "};\n"
|
res append "};\n"
|
||||||
res append "core {\n"
|
res append "core {\n"
|
||||||
for (i <- 0 until site(NTiles)) {
|
for (i <- 0 until site(NTiles)) {
|
||||||
val timecmpAddr = rtcAddr + 8*(i+1)
|
val isa = s"rv${site(XLen)}ima${if (site(UseFPU)) "fd" else ""}"
|
||||||
|
val timecmpAddr = addrMap("io:int:rtc").start + 8*(i+1)
|
||||||
res append s" $i {\n"
|
res append s" $i {\n"
|
||||||
res append " 0 {\n"
|
res append " 0 {\n"
|
||||||
res append s" isa rv$xLen;\n"
|
res append s" isa $isa;\n"
|
||||||
res append s" timecmp 0x${timecmpAddr.toString(16)};\n"
|
res append s" timecmp 0x${timecmpAddr.toString(16)};\n"
|
||||||
res append " };\n"
|
res append " };\n"
|
||||||
res append " };\n"
|
res append " };\n"
|
||||||
@ -82,7 +94,7 @@ class DefaultConfig extends Config (
|
|||||||
// Bits needed at the L2 agent
|
// Bits needed at the L2 agent
|
||||||
log2Up(site(NAcquireTransactors)+2) +
|
log2Up(site(NAcquireTransactors)+2) +
|
||||||
// Bits added by NASTI interconnect
|
// Bits added by NASTI interconnect
|
||||||
log2Up(site(MaxBanksPerMemoryChannel)))
|
log2Up(site(NBanksPerMemoryChannel)))
|
||||||
case MIFDataBits => Dump("MIF_DATA_BITS", 64)
|
case MIFDataBits => Dump("MIF_DATA_BITS", 64)
|
||||||
case MIFAddrBits => Dump("MIF_ADDR_BITS",
|
case MIFAddrBits => Dump("MIF_ADDR_BITS",
|
||||||
site(PAddrBits) - site(CacheBlockOffsetBits))
|
site(PAddrBits) - site(CacheBlockOffsetBits))
|
||||||
@ -153,7 +165,6 @@ class DefaultConfig extends Config (
|
|||||||
case RoccNMemChannels => site(BuildRoCC).map(_.nMemChannels).foldLeft(0)(_ + _)
|
case RoccNMemChannels => site(BuildRoCC).map(_.nMemChannels).foldLeft(0)(_ + _)
|
||||||
case RoccNPTWPorts => site(BuildRoCC).map(_.nPTWPorts).foldLeft(0)(_ + _)
|
case RoccNPTWPorts => site(BuildRoCC).map(_.nPTWPorts).foldLeft(0)(_ + _)
|
||||||
case RoccNCSRs => site(BuildRoCC).map(_.csrs.size).foldLeft(0)(_ + _)
|
case RoccNCSRs => site(BuildRoCC).map(_.csrs.size).foldLeft(0)(_ + _)
|
||||||
case UseStreamLoopback => false
|
|
||||||
case NDmaTransactors => 3
|
case NDmaTransactors => 3
|
||||||
case NDmaXacts => site(NDmaTransactors) * site(NTiles)
|
case NDmaXacts => site(NDmaTransactors) * site(NTiles)
|
||||||
case NDmaClients => site(NTiles)
|
case NDmaClients => site(NTiles)
|
||||||
@ -217,7 +228,7 @@ class DefaultConfig extends Config (
|
|||||||
dataBits = site(CacheBlockBytes)*8)
|
dataBits = site(CacheBlockBytes)*8)
|
||||||
case TLKey("Outermost") => site(TLKey("L2toMC")).copy(
|
case TLKey("Outermost") => site(TLKey("L2toMC")).copy(
|
||||||
maxClientXacts = site(NAcquireTransactors) + 2,
|
maxClientXacts = site(NAcquireTransactors) + 2,
|
||||||
maxClientsPerPort = site(MaxBanksPerMemoryChannel),
|
maxClientsPerPort = site(NBanksPerMemoryChannel),
|
||||||
dataBeats = site(MIFDataBeats))
|
dataBeats = site(MIFDataBeats))
|
||||||
case TLKey("L2toMMIO") => {
|
case TLKey("L2toMMIO") => {
|
||||||
val addrMap = new AddrHashMap(site(GlobalAddrMap))
|
val addrMap = new AddrHashMap(site(GlobalAddrMap))
|
||||||
@ -236,29 +247,13 @@ class DefaultConfig extends Config (
|
|||||||
case NTiles => Knob("NTILES")
|
case NTiles => Knob("NTILES")
|
||||||
case NMemoryChannels => Dump("N_MEM_CHANNELS", 1)
|
case NMemoryChannels => Dump("N_MEM_CHANNELS", 1)
|
||||||
case NBanksPerMemoryChannel => Knob("NBANKS_PER_MEM_CHANNEL")
|
case NBanksPerMemoryChannel => Knob("NBANKS_PER_MEM_CHANNEL")
|
||||||
case MemoryChannelMuxConfigs => Dump("MEMORY_CHANNEL_MUX_CONFIGS", List( site(NMemoryChannels) ))
|
case NOutstandingMemReqsPerChannel => site(NBanksPerMemoryChannel)*(site(NAcquireTransactors)+2)
|
||||||
case MaxBanksPerMemoryChannel => site(NBanksPerMemoryChannel) * site(NMemoryChannels) / site(MemoryChannelMuxConfigs).sortWith{_ < _}(0)
|
|
||||||
case NOutstandingMemReqsPerChannel => site(MaxBanksPerMemoryChannel)*(site(NAcquireTransactors)+2)
|
|
||||||
case BankIdLSB => 0
|
case BankIdLSB => 0
|
||||||
case CacheBlockBytes => Dump("CACHE_BLOCK_BYTES", 64)
|
case CacheBlockBytes => Dump("CACHE_BLOCK_BYTES", 64)
|
||||||
case CacheBlockOffsetBits => log2Up(here(CacheBlockBytes))
|
case CacheBlockOffsetBits => log2Up(here(CacheBlockBytes))
|
||||||
case UseHtifClockDiv => true
|
case UseHtifClockDiv => true
|
||||||
case ConfigString => makeConfigString()
|
case ConfigString => makeConfigString()
|
||||||
case GlobalAddrMap => {
|
case GlobalAddrMap => globalAddrMap
|
||||||
val memsize = BigInt(1L << 30)
|
|
||||||
Dump("MEM_SIZE", memsize)
|
|
||||||
AddrMap(
|
|
||||||
AddrMapEntry("mem", MemSize(memsize, AddrMapConsts.RWX, true)),
|
|
||||||
AddrMapEntry("conf", MemSubmap(BigInt(1L << 30), genCsrAddrMap)),
|
|
||||||
AddrMapEntry("devices", MemSubmap(BigInt(1L << 31), site(GlobalDeviceSet).getAddrMap)))
|
|
||||||
}
|
|
||||||
case GlobalDeviceSet => {
|
|
||||||
val devset = new DeviceSet
|
|
||||||
if (site(UseStreamLoopback)) {
|
|
||||||
devset.addDevice("loopback", site(StreamLoopbackWidth) / 8, "stream")
|
|
||||||
}
|
|
||||||
devset
|
|
||||||
}
|
|
||||||
case _ => throw new CDEMatchError
|
case _ => throw new CDEMatchError
|
||||||
}},
|
}},
|
||||||
knobValues = {
|
knobValues = {
|
||||||
@ -488,19 +483,6 @@ class OctoChannelBenchmarkConfig extends Config(new With8MemoryChannels ++ new S
|
|||||||
|
|
||||||
class EightChannelVLSIConfig extends Config(new With8MemoryChannels ++ new DefaultVLSIConfig)
|
class EightChannelVLSIConfig extends Config(new With8MemoryChannels ++ new DefaultVLSIConfig)
|
||||||
|
|
||||||
class WithOneOrMaxChannels extends Config(
|
|
||||||
(pname, site, here) => pname match {
|
|
||||||
case MemoryChannelMuxConfigs => Dump("MEMORY_CHANNEL_MUX_CONFIGS", List(1, site(NMemoryChannels)))
|
|
||||||
case _ => throw new CDEMatchError
|
|
||||||
}
|
|
||||||
)
|
|
||||||
class OneOrEightChannelBenchmarkConfig extends Config(new WithOneOrMaxChannels ++ new With8MemoryChannels ++ new SingleChannelBenchmarkConfig)
|
|
||||||
class OneOrEightChannelVLSIConfig extends Config(new WithOneOrMaxChannels ++ new EightChannelVLSIConfig)
|
|
||||||
|
|
||||||
class SimulateBackupMemConfig extends Config(){ Dump("MEM_BACKUP_EN", true) }
|
|
||||||
class BackupMemVLSIConfig extends Config(new SimulateBackupMemConfig ++ new DefaultVLSIConfig)
|
|
||||||
class OneOrEightChannelBackupMemVLSIConfig extends Config(new WithOneOrMaxChannels ++ new With8MemoryChannels ++ new BackupMemVLSIConfig)
|
|
||||||
|
|
||||||
class WithSplitL2Metadata extends Config(knobValues = { case "L2_SPLIT_METADATA" => true; case _ => throw new CDEMatchError })
|
class WithSplitL2Metadata extends Config(knobValues = { case "L2_SPLIT_METADATA" => true; case _ => throw new CDEMatchError })
|
||||||
class SplitL2MetadataTestConfig extends Config(new WithSplitL2Metadata ++ new DefaultL2Config)
|
class SplitL2MetadataTestConfig extends Config(new WithSplitL2Metadata ++ new DefaultL2Config)
|
||||||
|
|
||||||
|
@ -4,7 +4,6 @@ import Chisel.log2Ceil
|
|||||||
import cde.{Parameters, Field}
|
import cde.{Parameters, Field}
|
||||||
import scala.collection.mutable.HashMap
|
import scala.collection.mutable.HashMap
|
||||||
import junctions._
|
import junctions._
|
||||||
import junctions.AddrMapConsts._
|
|
||||||
|
|
||||||
case object GlobalDeviceSet extends Field[DeviceSet]
|
case object GlobalDeviceSet extends Field[DeviceSet]
|
||||||
|
|
||||||
@ -27,9 +26,9 @@ class DeviceSet {
|
|||||||
def getAddrMap: AddrMap = {
|
def getAddrMap: AddrMap = {
|
||||||
val devices = this.toSeq.sortWith((a, b) => a.size > b.size)
|
val devices = this.toSeq.sortWith((a, b) => a.size > b.size)
|
||||||
val entries = devices.map { case Device(name, size, _, readable, writeable) =>
|
val entries = devices.map { case Device(name, size, _, readable, writeable) =>
|
||||||
val prot = (if (readable) R else 0) | (if (writeable) W else 0)
|
val prot = (if (readable) AddrMapProt.R else 0) | (if (writeable) AddrMapProt.W else 0)
|
||||||
val realsize = roundup(size)
|
val realsize = roundup(size)
|
||||||
new AddrMapEntry(name, new MemSize(realsize, prot))
|
AddrMapEntry(name, MemSize(size, roundup(size), MemAttr(prot)))
|
||||||
}
|
}
|
||||||
new AddrMap(entries)
|
new AddrMap(entries)
|
||||||
}
|
}
|
||||||
|
@ -17,10 +17,6 @@ case object NTiles extends Field[Int]
|
|||||||
case object NMemoryChannels extends Field[Int]
|
case object NMemoryChannels extends Field[Int]
|
||||||
/** Number of banks per memory channel */
|
/** Number of banks per memory channel */
|
||||||
case object NBanksPerMemoryChannel extends Field[Int]
|
case object NBanksPerMemoryChannel extends Field[Int]
|
||||||
/** Maximum number of banks per memory channel, when configurable */
|
|
||||||
case object MaxBanksPerMemoryChannel extends Field[Int]
|
|
||||||
/** Dynamic memory channel configurations */
|
|
||||||
case object MemoryChannelMuxConfigs extends Field[List[Int]]
|
|
||||||
/** Least significant bit of address used for bank partitioning */
|
/** Least significant bit of address used for bank partitioning */
|
||||||
case object BankIdLSB extends Field[Int]
|
case object BankIdLSB extends Field[Int]
|
||||||
/** Number of outstanding memory requests */
|
/** Number of outstanding memory requests */
|
||||||
@ -61,7 +57,6 @@ trait HasTopLevelParameters {
|
|||||||
lazy val scrAddrBits = log2Up(nSCR)
|
lazy val scrAddrBits = log2Up(nSCR)
|
||||||
lazy val scrDataBits = 64
|
lazy val scrDataBits = 64
|
||||||
lazy val scrDataBytes = scrDataBits / 8
|
lazy val scrDataBytes = scrDataBits / 8
|
||||||
lazy val memoryChannelMuxConfigs = p(MemoryChannelMuxConfigs)
|
|
||||||
//require(lsb + log2Up(nBanks) < mifAddrBits)
|
//require(lsb + log2Up(nBanks) < mifAddrBits)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -80,6 +75,7 @@ class BasicTopIO(implicit val p: Parameters) extends ParameterizedBundle()(p)
|
|||||||
|
|
||||||
class TopIO(implicit p: Parameters) extends BasicTopIO()(p) {
|
class TopIO(implicit p: Parameters) extends BasicTopIO()(p) {
|
||||||
val mem = Vec(nMemChannels, new NastiIO)
|
val mem = Vec(nMemChannels, new NastiIO)
|
||||||
|
val mmio = new NastiIO
|
||||||
}
|
}
|
||||||
|
|
||||||
object TopUtils {
|
object TopUtils {
|
||||||
@ -125,9 +121,11 @@ class Top(topParams: Parameters) extends Module with HasTopLevelParameters {
|
|||||||
uncore.io.tiles_uncached <> tileList.map(_.io.uncached).flatten
|
uncore.io.tiles_uncached <> tileList.map(_.io.uncached).flatten
|
||||||
io.host <> uncore.io.host
|
io.host <> uncore.io.host
|
||||||
|
|
||||||
|
io.mmio <> uncore.io.mmio
|
||||||
io.mem.zip(uncore.io.mem).foreach { case (outer, inner) =>
|
io.mem.zip(uncore.io.mem).foreach { case (outer, inner) =>
|
||||||
TopUtils.connectNasti(outer, inner)
|
outer <> inner
|
||||||
// Memory cache type should be normal non-cacheable bufferable
|
// Memory cache type should be normal non-cacheable bufferable
|
||||||
|
// TODO why is this happening here? Would 0000 (device) be OK instead?
|
||||||
outer.ar.bits.cache := UInt("b0011")
|
outer.ar.bits.cache := UInt("b0011")
|
||||||
outer.aw.bits.cache := UInt("b0011")
|
outer.aw.bits.cache := UInt("b0011")
|
||||||
}
|
}
|
||||||
@ -165,23 +163,14 @@ class Uncore(implicit val p: Parameters) extends Module
|
|||||||
|
|
||||||
val addrMap = p(GlobalAddrMap)
|
val addrMap = p(GlobalAddrMap)
|
||||||
val addrHashMap = new AddrHashMap(addrMap)
|
val addrHashMap = new AddrHashMap(addrMap)
|
||||||
val memSize = addrHashMap("mem").size
|
|
||||||
val scrFile = Module(new SCRFile("UNCORE_SCR", 0))
|
val scrFile = Module(new SCRFile("UNCORE_SCR", 0))
|
||||||
scrFile.io.smi <> htif.io.scr
|
scrFile.io.smi <> htif.io.scr
|
||||||
scrFile.io.scr.attach(Wire(init = UInt(nTiles)), "N_CORES")
|
scrFile.io.scr.attach(Wire(init = UInt(nTiles)), "N_CORES")
|
||||||
scrFile.io.scr.attach(Wire(init = UInt(memSize >> 20)), "MMIO_BASE")
|
scrFile.io.scr.attach(Wire(init = UInt(addrHashMap("io:int:configstring").start >> 20)), "MMIO_BASE")
|
||||||
// scrFile.io.scr <> (... your SCR connections ...)
|
// scrFile.io.scr <> (... your SCR connections ...)
|
||||||
|
|
||||||
buildMMIONetwork(p.alterPartial({case TLId => "MMIO_Outermost"}))
|
buildMMIONetwork(p.alterPartial({case TLId => "MMIO_Outermost"}))
|
||||||
|
|
||||||
// Configures the enabled memory channels. This can't be changed while the
|
|
||||||
// chip is actively using memory, as it both drops Nasti messages and garbles
|
|
||||||
// all of memory.
|
|
||||||
val memory_channel_mux_select = scrFile.io.scr.attach(
|
|
||||||
Reg(UInt(width = log2Up(memoryChannelMuxConfigs.size))),
|
|
||||||
"MEMORY_CHANNEL_MUX_SELECT")
|
|
||||||
outmemsys.io.memory_channel_mux_select := memory_channel_mux_select
|
|
||||||
|
|
||||||
// Wire the htif to the memory port(s) and host interface
|
// Wire the htif to the memory port(s) and host interface
|
||||||
io.mem <> outmemsys.io.mem
|
io.mem <> outmemsys.io.mem
|
||||||
if(p(UseHtifClockDiv)) {
|
if(p(UseHtifClockDiv)) {
|
||||||
@ -191,31 +180,26 @@ class Uncore(implicit val p: Parameters) extends Module
|
|||||||
}
|
}
|
||||||
|
|
||||||
def buildMMIONetwork(implicit p: Parameters) = {
|
def buildMMIONetwork(implicit p: Parameters) = {
|
||||||
|
val (ioBase, ioAddrMap) = addrHashMap.subMap("io")
|
||||||
|
val ioAddrHashMap = new AddrHashMap(ioAddrMap, ioBase)
|
||||||
|
|
||||||
val mmioNarrower = Module(new TileLinkIONarrower("L2toMMIO", "MMIO_Outermost"))
|
val mmioNarrower = Module(new TileLinkIONarrower("L2toMMIO", "MMIO_Outermost"))
|
||||||
val mmioNetwork = Module(new TileLinkRecursiveInterconnect(1, addrMap.tail, memSize))
|
val mmioNetwork = Module(new TileLinkRecursiveInterconnect(1, ioAddrMap, ioBase))
|
||||||
|
|
||||||
mmioNarrower.io.in <> outmemsys.io.mmio
|
mmioNarrower.io.in <> outmemsys.io.mmio
|
||||||
mmioNetwork.io.in.head <> mmioNarrower.io.out
|
mmioNetwork.io.in.head <> mmioNarrower.io.out
|
||||||
|
|
||||||
if (p(UseStreamLoopback)) {
|
|
||||||
val lo_width = p(StreamLoopbackWidth)
|
|
||||||
val lo_size = p(StreamLoopbackSize)
|
|
||||||
val lo_conv = Module(new NastiIOStreamIOConverter(lo_width))
|
|
||||||
val lo_port = addrHashMap("devices:loopback").port - 1
|
|
||||||
TopUtils.connectTilelinkNasti(lo_conv.io.nasti, mmioNetwork.io.out(lo_port))
|
|
||||||
lo_conv.io.stream.in <> Queue(lo_conv.io.stream.out, lo_size)
|
|
||||||
}
|
|
||||||
|
|
||||||
val rtc = Module(new RTC(p(NTiles)))
|
val rtc = Module(new RTC(p(NTiles)))
|
||||||
val rtcAddr = addrHashMap("conf:rtc")
|
val rtcAddr = ioAddrHashMap("int:rtc")
|
||||||
val rtcPort = rtcAddr.port - 1
|
require(rtc.size <= rtcAddr.region.size)
|
||||||
require(rtc.size <= rtcAddr.size)
|
rtc.io.tl <> mmioNetwork.io.out(rtcAddr.port)
|
||||||
rtc.io.tl <> mmioNetwork.io.out(rtcPort)
|
|
||||||
io.timerIRQs := rtc.io.irqs
|
io.timerIRQs := rtc.io.irqs
|
||||||
|
|
||||||
val deviceTree = Module(new ROMSlave(p(ConfigString).toSeq))
|
val deviceTree = Module(new ROMSlave(p(ConfigString).toSeq))
|
||||||
val dtPort = addrHashMap("conf:devicetree").port - 1
|
val deviceTreeAddr = ioAddrHashMap("int:configstring")
|
||||||
deviceTree.io <> mmioNetwork.io.out(dtPort)
|
deviceTree.io <> mmioNetwork.io.out(deviceTreeAddr.port)
|
||||||
|
|
||||||
|
TopUtils.connectTilelinkNasti(io.mmio, mmioNetwork.io.out(ioAddrHashMap("ext").port))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -229,7 +213,6 @@ class OuterMemorySystem(implicit val p: Parameters) extends Module with HasTopLe
|
|||||||
val htif_uncached = (new ClientUncachedTileLinkIO).flip
|
val htif_uncached = (new ClientUncachedTileLinkIO).flip
|
||||||
val incoherent = Vec(nTiles, Bool()).asInput
|
val incoherent = Vec(nTiles, Bool()).asInput
|
||||||
val mem = Vec(nMemChannels, new NastiIO)
|
val mem = Vec(nMemChannels, new NastiIO)
|
||||||
val memory_channel_mux_select = UInt(INPUT, log2Up(memoryChannelMuxConfigs.size))
|
|
||||||
val mmio = new ClientUncachedTileLinkIO()(p.alterPartial({case TLId => "L2toMMIO"}))
|
val mmio = new ClientUncachedTileLinkIO()(p.alterPartial({case TLId => "L2toMMIO"}))
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -272,34 +255,20 @@ class OuterMemorySystem(implicit val p: Parameters) extends Module with HasTopLe
|
|||||||
|
|
||||||
// TODO: the code to print this stuff should live somewhere else
|
// TODO: the code to print this stuff should live somewhere else
|
||||||
println("Generated Address Map")
|
println("Generated Address Map")
|
||||||
for ((name, base, size, _, _) <- addrHashMap.sortedEntries) {
|
for ((name, base, region) <- addrHashMap.sortedEntries) {
|
||||||
println(f"\t$name%s $base%x - ${base + size - 1}%x")
|
println(f"\t$name%s $base%x - ${base + region.size - 1}%x")
|
||||||
}
|
}
|
||||||
println("Generated Configuration String")
|
println("Generated Configuration String")
|
||||||
println(new String(p(ConfigString)))
|
println(new String(p(ConfigString)))
|
||||||
|
|
||||||
val channelConfigs = p(MemoryChannelMuxConfigs)
|
val mem_ic = Module(new TileLinkMemoryInterconnect(nBanksPerMemChannel, nMemChannels)(outermostTLParams))
|
||||||
require(channelConfigs.sortWith(_ > _)(0) == nMemChannels,
|
|
||||||
"More memory channels elaborated than can be enabled")
|
|
||||||
val mem_ic =
|
|
||||||
if (channelConfigs.size == 1) {
|
|
||||||
val ic = Module(new TileLinkMemoryInterconnect(
|
|
||||||
nBanksPerMemChannel, nMemChannels)(outermostTLParams))
|
|
||||||
ic
|
|
||||||
} else {
|
|
||||||
val nBanks = nBanksPerMemChannel * nMemChannels
|
|
||||||
val ic = Module(new TileLinkMemorySelector(
|
|
||||||
nBanks, nMemChannels, channelConfigs)(outermostTLParams))
|
|
||||||
ic.io.select := io.memory_channel_mux_select
|
|
||||||
ic
|
|
||||||
}
|
|
||||||
|
|
||||||
for ((bank, i) <- managerEndpoints.zipWithIndex) {
|
for ((bank, icPort) <- managerEndpoints zip mem_ic.io.in) {
|
||||||
val unwrap = Module(new ClientTileLinkIOUnwrapper()(outerTLParams))
|
val unwrap = Module(new ClientTileLinkIOUnwrapper()(outerTLParams))
|
||||||
val narrow = Module(new TileLinkIONarrower("L2toMC", "Outermost"))
|
val narrow = Module(new TileLinkIONarrower("L2toMC", "Outermost"))
|
||||||
unwrap.io.in <> ClientTileLinkEnqueuer(bank.outerTL, backendBuffering)(outerTLParams)
|
unwrap.io.in <> ClientTileLinkEnqueuer(bank.outerTL, backendBuffering)(outerTLParams)
|
||||||
narrow.io.in <> unwrap.io.out
|
narrow.io.in <> unwrap.io.out
|
||||||
mem_ic.io.in(i) <> narrow.io.out
|
icPort <> narrow.io.out
|
||||||
}
|
}
|
||||||
|
|
||||||
for ((nasti, tl) <- io.mem zip mem_ic.io.out)
|
for ((nasti, tl) <- io.mem zip mem_ic.io.out)
|
||||||
|
@ -155,7 +155,3 @@ class TraceGenConfig extends Config(new With2Cores ++ new WithL2Cache ++ new Wit
|
|||||||
class FancyMemtestConfig extends Config(
|
class FancyMemtestConfig extends Config(
|
||||||
new With2Cores ++ new With2MemoryChannels ++ new With4BanksPerMemChannel ++
|
new With2Cores ++ new With2MemoryChannels ++ new With4BanksPerMemChannel ++
|
||||||
new WithMemtest ++ new WithL2Cache ++ new GroundTestConfig)
|
new WithMemtest ++ new WithL2Cache ++ new GroundTestConfig)
|
||||||
|
|
||||||
class MemoryMuxMemtestConfig extends Config(
|
|
||||||
new With2MemoryChannels ++ new WithOneOrMaxChannels ++
|
|
||||||
new WithMemtest ++ new GroundTestConfig)
|
|
||||||
|
2
uncore
2
uncore
@ -1 +1 @@
|
|||||||
Subproject commit 291a0bbf58006bc1d2587c58fa7d63abd7d57083
|
Subproject commit c57928fad82beb9e7cab82cb03cf394b24512351
|
Loading…
Reference in New Issue
Block a user