236 lines
6.9 KiB
Scala
236 lines
6.9 KiB
Scala
package rocket
|
|
|
|
import Chisel._;
|
|
import Node._;
|
|
import Constants._;
|
|
import scala.math._;
|
|
|
|
class ioDmemArbiter(n: Int) extends Bundle
|
|
{
|
|
val dmem = new ioDmem().flip
|
|
val requestor = Vec(n) { new ioDmem() }
|
|
}
|
|
|
|
class rocketDmemArbiter(n: Int) extends Component
|
|
{
|
|
val io = new ioDmemArbiter(n)
|
|
require(DCACHE_TAG_BITS >= log2up(n) + CPU_TAG_BITS)
|
|
|
|
var req_val = Bool(false)
|
|
var req_rdy = io.dmem.req_rdy
|
|
for (i <- 0 until n)
|
|
{
|
|
io.requestor(i).req_rdy := req_rdy
|
|
req_val = req_val || io.requestor(i).req_val
|
|
req_rdy = req_rdy && !io.requestor(i).req_val
|
|
}
|
|
|
|
var req_cmd = io.requestor(n-1).req_cmd
|
|
var req_type = io.requestor(n-1).req_type
|
|
var req_idx = io.requestor(n-1).req_idx
|
|
var req_ppn = io.requestor(n-1).req_ppn
|
|
var req_data = io.requestor(n-1).req_data
|
|
var req_tag = io.requestor(n-1).req_tag
|
|
var req_kill = io.requestor(n-1).req_kill
|
|
for (i <- n-1 to 0 by -1)
|
|
{
|
|
req_cmd = Mux(io.requestor(i).req_val, io.requestor(i).req_cmd, req_cmd)
|
|
req_type = Mux(io.requestor(i).req_val, io.requestor(i).req_type, req_type)
|
|
req_idx = Mux(io.requestor(i).req_val, io.requestor(i).req_idx, req_idx)
|
|
req_ppn = Mux(Reg(io.requestor(i).req_val), io.requestor(i).req_ppn, req_ppn)
|
|
req_data = Mux(Reg(io.requestor(i).req_val), io.requestor(i).req_data, req_data)
|
|
req_tag = Mux(io.requestor(i).req_val, Cat(io.requestor(i).req_tag, UFix(i, log2up(n))), req_tag)
|
|
req_kill = Mux(Reg(io.requestor(i).req_val), io.requestor(i).req_kill, req_kill)
|
|
}
|
|
|
|
io.dmem.req_val := req_val
|
|
io.dmem.req_cmd := req_cmd
|
|
io.dmem.req_type := req_type
|
|
io.dmem.req_idx := req_idx
|
|
io.dmem.req_ppn := req_ppn
|
|
io.dmem.req_data := req_data
|
|
io.dmem.req_tag := req_tag
|
|
io.dmem.req_kill := req_kill
|
|
|
|
for (i <- 0 until n)
|
|
{
|
|
val tag_hit = io.dmem.resp_tag(log2up(n)-1,0) === UFix(i)
|
|
io.requestor(i).xcpt_ma_ld := io.dmem.xcpt_ma_ld && Reg(io.requestor(i).req_val)
|
|
io.requestor(i).xcpt_ma_st := io.dmem.xcpt_ma_st && Reg(io.requestor(i).req_val)
|
|
io.requestor(i).resp_nack := io.dmem.resp_nack && Reg(io.requestor(i).req_val)
|
|
io.requestor(i).resp_miss := io.dmem.resp_miss && tag_hit
|
|
io.requestor(i).resp_val := io.dmem.resp_val && tag_hit
|
|
io.requestor(i).resp_replay := io.dmem.resp_replay && tag_hit
|
|
io.requestor(i).resp_data := io.dmem.resp_data
|
|
io.requestor(i).resp_data_subword := io.dmem.resp_data_subword
|
|
io.requestor(i).resp_type := io.dmem.resp_type
|
|
io.requestor(i).resp_tag := io.dmem.resp_tag >> UFix(log2up(n))
|
|
}
|
|
}
|
|
|
|
class ioPTW extends Bundle
|
|
{
|
|
val itlb = new ioTLB_PTW().flip
|
|
val dtlb = new ioTLB_PTW().flip
|
|
val dmem = new ioDmem().flip
|
|
val ptbr = UFix(PADDR_BITS, INPUT);
|
|
}
|
|
|
|
class rocketPTW extends Component
|
|
{
|
|
val io = new ioPTW();
|
|
|
|
val s_ready :: s_l1_req :: s_l1_wait :: s_l1_fake :: s_l2_req :: s_l2_wait :: s_l2_fake:: s_l3_req :: s_l3_wait :: s_done :: s_error :: Nil = Enum(11) { UFix() };
|
|
val state = Reg(resetVal = s_ready);
|
|
|
|
val r_req_vpn = Reg() { Bits() }
|
|
val r_req_dest = Reg() { Bool() }
|
|
|
|
val req_addr = Reg() { UFix() };
|
|
val r_resp_ppn = Reg() { Bits() };
|
|
val r_resp_perm = Reg() { Bits() };
|
|
|
|
val vpn_idx = Mux(state === s_l2_wait, r_req_vpn(9,0), r_req_vpn(19,10));
|
|
val req_val = io.itlb.req_val || io.dtlb.req_val;
|
|
|
|
// give ITLB requests priority over DTLB requests
|
|
val req_itlb_val = io.itlb.req_val;
|
|
val req_dtlb_val = io.dtlb.req_val && !io.itlb.req_val;
|
|
|
|
when ((state === s_ready) && req_dtlb_val) {
|
|
r_req_vpn := io.dtlb.req_vpn;
|
|
r_req_dest := Bool(true);
|
|
req_addr := Cat(io.ptbr(PADDR_BITS-1,PGIDX_BITS), io.dtlb.req_vpn(VPN_BITS-1,VPN_BITS-10), Bits(0,3)).toUFix;
|
|
}
|
|
|
|
when ((state === s_ready) && req_itlb_val) {
|
|
r_req_vpn := io.itlb.req_vpn;
|
|
r_req_dest := Bool(false);
|
|
req_addr := Cat(io.ptbr(PADDR_BITS-1,PGIDX_BITS), io.itlb.req_vpn(VPN_BITS-1,VPN_BITS-10), Bits(0,3)).toUFix;
|
|
}
|
|
|
|
when (io.dmem.resp_val) {
|
|
req_addr := Cat(io.dmem.resp_data(PADDR_BITS-1, PGIDX_BITS), vpn_idx, Bits(0,3)).toUFix;
|
|
r_resp_perm := io.dmem.resp_data(9,4);
|
|
r_resp_ppn := io.dmem.resp_data(PADDR_BITS-1, PGIDX_BITS);
|
|
}
|
|
|
|
io.dmem.req_val :=
|
|
(state === s_l1_req) ||
|
|
(state === s_l2_req) ||
|
|
(state === s_l3_req);
|
|
|
|
io.dmem.req_cmd := M_XRD;
|
|
io.dmem.req_type := MT_D;
|
|
io.dmem.req_idx := req_addr(PGIDX_BITS-1,0);
|
|
io.dmem.req_ppn := Reg(req_addr(PADDR_BITS-1,PGIDX_BITS))
|
|
io.dmem.req_kill := Bool(false)
|
|
|
|
val resp_val = (state === s_done) || (state === s_l1_fake) || (state === s_l2_fake);
|
|
val resp_err = (state === s_error);
|
|
|
|
val resp_ptd = (io.dmem.resp_data(1,0) === Bits(1,2));
|
|
val resp_pte = (io.dmem.resp_data(1,0) === Bits(2,2));
|
|
|
|
io.dtlb.req_rdy := (state === s_ready) && !io.itlb.req_val;
|
|
io.itlb.req_rdy := (state === s_ready);
|
|
io.dtlb.resp_val := r_req_dest && resp_val;
|
|
io.itlb.resp_val := !r_req_dest && resp_val;
|
|
io.dtlb.resp_err := r_req_dest && resp_err;
|
|
io.itlb.resp_err := !r_req_dest && resp_err;
|
|
io.dtlb.resp_perm := r_resp_perm;
|
|
io.itlb.resp_perm := r_resp_perm;
|
|
|
|
val resp_ppn =
|
|
Mux(state === s_l1_fake, Cat(r_resp_ppn(PPN_BITS-1, PPN_BITS-7), r_req_vpn(VPN_BITS-11, 0)),
|
|
Mux(state === s_l2_fake, Cat(r_resp_ppn(PPN_BITS-1, PPN_BITS-17), r_req_vpn(VPN_BITS-21, 0)),
|
|
r_resp_ppn));
|
|
|
|
io.dtlb.resp_ppn := resp_ppn;
|
|
io.itlb.resp_ppn := resp_ppn;
|
|
|
|
// control state machine
|
|
switch (state) {
|
|
is (s_ready) {
|
|
when (req_val) {
|
|
state := s_l1_req;
|
|
}
|
|
}
|
|
// level 1
|
|
is (s_l1_req) {
|
|
when (io.dmem.req_rdy) {
|
|
state := s_l1_wait;
|
|
}
|
|
}
|
|
is (s_l1_wait) {
|
|
when (io.dmem.resp_nack) {
|
|
state := s_l1_req
|
|
}
|
|
when (io.dmem.resp_val) {
|
|
when (resp_ptd) { // page table descriptor
|
|
state := s_l2_req;
|
|
}
|
|
.elsewhen (resp_pte) { // page table entry
|
|
state := s_l1_fake;
|
|
}
|
|
.otherwise {
|
|
state := s_error;
|
|
}
|
|
}
|
|
}
|
|
is (s_l1_fake) {
|
|
state := s_ready;
|
|
}
|
|
// level 2
|
|
is (s_l2_req) {
|
|
when (io.dmem.req_rdy) {
|
|
state := s_l2_wait;
|
|
}
|
|
}
|
|
is (s_l2_wait) {
|
|
when (io.dmem.resp_nack) {
|
|
state := s_l2_req
|
|
}
|
|
when (io.dmem.resp_val) {
|
|
when (resp_ptd) { // page table descriptor
|
|
state := s_l3_req;
|
|
}
|
|
.elsewhen (resp_pte) { // page table entry
|
|
state := s_l2_fake;
|
|
}
|
|
.otherwise {
|
|
state := s_error;
|
|
}
|
|
}
|
|
}
|
|
is (s_l2_fake) {
|
|
state := s_ready;
|
|
}
|
|
// level 3
|
|
is (s_l3_req) {
|
|
when (io.dmem.req_rdy) {
|
|
state := s_l3_wait;
|
|
}
|
|
}
|
|
is (s_l3_wait) {
|
|
when (io.dmem.resp_nack) {
|
|
state := s_l3_req
|
|
}
|
|
when (io.dmem.resp_val) {
|
|
when (resp_pte) { // page table entry
|
|
state := s_done;
|
|
}
|
|
.otherwise {
|
|
state := s_error;
|
|
}
|
|
}
|
|
}
|
|
is (s_done) {
|
|
state := s_ready;
|
|
}
|
|
is (s_error) {
|
|
state := s_ready;
|
|
}
|
|
}
|
|
}
|