xref: /XiangShan/src/main/scala/xiangshan/cache/mmu/PageTableWalker.scala (revision 7797f0350058297bc0fbf659b4ba5606b33415e8)
16d5ddbceSLemover/***************************************************************************************
26d5ddbceSLemover* Copyright (c) 2020-2021 Institute of Computing Technology, Chinese Academy of Sciences
3f320e0f0SYinan Xu* Copyright (c) 2020-2021 Peng Cheng Laboratory
46d5ddbceSLemover*
56d5ddbceSLemover* XiangShan is licensed under Mulan PSL v2.
66d5ddbceSLemover* You can use this software according to the terms and conditions of the Mulan PSL v2.
76d5ddbceSLemover* You may obtain a copy of Mulan PSL v2 at:
86d5ddbceSLemover*          http://license.coscl.org.cn/MulanPSL2
96d5ddbceSLemover*
106d5ddbceSLemover* THIS SOFTWARE IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OF ANY KIND,
116d5ddbceSLemover* EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO NON-INFRINGEMENT,
126d5ddbceSLemover* MERCHANTABILITY OR FIT FOR A PARTICULAR PURPOSE.
136d5ddbceSLemover*
146d5ddbceSLemover* See the Mulan PSL v2 for more details.
156d5ddbceSLemover***************************************************************************************/
166d5ddbceSLemover
176d5ddbceSLemoverpackage xiangshan.cache.mmu
186d5ddbceSLemover
196d5ddbceSLemoverimport chipsalliance.rocketchip.config.Parameters
206d5ddbceSLemoverimport chisel3._
216d5ddbceSLemoverimport chisel3.util._
22b848eea5SLemoverimport chisel3.internal.naming.chiselName
236d5ddbceSLemoverimport xiangshan._
246d5ddbceSLemoverimport xiangshan.cache.{HasDCacheParameters, MemoryOpConstants}
256d5ddbceSLemoverimport utils._
266d5ddbceSLemoverimport freechips.rocketchip.diplomacy.{LazyModule, LazyModuleImp}
276d5ddbceSLemoverimport freechips.rocketchip.tilelink._
28b6982e83SLemoverimport xiangshan.backend.fu.{PMPReqBundle, PMPRespBundle}
296d5ddbceSLemover
3092e3bfefSLemover/** Page Table Walk is divided into two parts
3192e3bfefSLemover  * One,   PTW: page walk for pde, except for leaf entries, one by one
3292e3bfefSLemover  * Two, LLPTW: page walk for pte, only the leaf entries(4KB), in parallel
336d5ddbceSLemover  */
3492e3bfefSLemover
3592e3bfefSLemover
3692e3bfefSLemover/** PTW : page table walker
3792e3bfefSLemover  * a finite state machine
3892e3bfefSLemover  * only take 1GB and 2MB page walks
3992e3bfefSLemover  * or in other words, except the last level(leaf)
4092e3bfefSLemover  **/
4192e3bfefSLemoverclass PTWIO()(implicit p: Parameters) extends MMUIOBaseBundle with HasPtwConst {
426d5ddbceSLemover  val req = Flipped(DecoupledIO(new Bundle {
4345f497a4Shappy-lx    val req_info = new L2TlbInnerBundle()
446d5ddbceSLemover    val l1Hit = Bool()
456d5ddbceSLemover    val ppn = UInt(ppnLen.W)
466d5ddbceSLemover  }))
476d5ddbceSLemover  val resp = DecoupledIO(new Bundle {
48bc063562SLemover    val source = UInt(bSourceWidth.W)
496d5ddbceSLemover    val resp = new PtwResp
506d5ddbceSLemover  })
516d5ddbceSLemover
5292e3bfefSLemover  val llptw = DecoupledIO(new LLPTWInBundle())
539c503409SLemover  // NOTE: llptw change from "connect to llptw" to "connect to page cache"
549c503409SLemover  // to avoid corner case that caused duplicate entries
55cc5a5f22SLemover
566d5ddbceSLemover  val mem = new Bundle {
57b848eea5SLemover    val req = DecoupledIO(new L2TlbMemReqBundle())
585854c1edSLemover    val resp = Flipped(ValidIO(UInt(XLEN.W)))
59cc5a5f22SLemover    val mask = Input(Bool())
606d5ddbceSLemover  }
61b6982e83SLemover  val pmp = new Bundle {
62b6982e83SLemover    val req = ValidIO(new PMPReqBundle())
63b6982e83SLemover    val resp = Flipped(new PMPRespBundle())
64b6982e83SLemover  }
656d5ddbceSLemover
666d5ddbceSLemover  val refill = Output(new Bundle {
6745f497a4Shappy-lx    val req_info = new L2TlbInnerBundle()
686d5ddbceSLemover    val level = UInt(log2Up(Level).W)
696d5ddbceSLemover  })
706d5ddbceSLemover}
716d5ddbceSLemover
72b848eea5SLemover@chiselName
7392e3bfefSLemoverclass PTW()(implicit p: Parameters) extends XSModule with HasPtwConst with HasPerfEvents {
7492e3bfefSLemover  val io = IO(new PTWIO)
756d5ddbceSLemover
766d5ddbceSLemover  val sfence = io.sfence
776d5ddbceSLemover  val mem = io.mem
786d5ddbceSLemover  val satp = io.csr.satp
7945f497a4Shappy-lx  val flush = io.sfence.valid || io.csr.satp.changed
806d5ddbceSLemover
81b6982e83SLemover  val s_idle :: s_addr_check :: s_mem_req :: s_mem_resp :: s_check_pte :: Nil = Enum(5)
826d5ddbceSLemover  val state = RegInit(s_idle)
836d5ddbceSLemover  val level = RegInit(0.U(log2Up(Level).W))
84b6982e83SLemover  val af_level = RegInit(0.U(log2Up(Level).W)) // access fault return this level
856d5ddbceSLemover  val ppn = Reg(UInt(ppnLen.W))
866d5ddbceSLemover  val vpn = Reg(UInt(vpnLen.W))
876d5ddbceSLemover  val levelNext = level + 1.U
886d5ddbceSLemover  val l1Hit = Reg(Bool())
895854c1edSLemover  val memPte = mem.resp.bits.asTypeOf(new PteBundle().cloneType)
90cc5a5f22SLemover  io.req.ready := state === s_idle
916d5ddbceSLemover
92b6982e83SLemover  val finish = WireInit(false.B)
93b6982e83SLemover  val sent_to_pmp = state === s_addr_check || (state === s_check_pte && !finish)
94ca2f90a6SLemover  val accessFault = RegEnable(io.pmp.resp.ld || io.pmp.resp.mmio, sent_to_pmp)
95b6982e83SLemover  val pageFault = memPte.isPf(level)
966d5ddbceSLemover  switch (state) {
976d5ddbceSLemover    is (s_idle) {
986d5ddbceSLemover      when (io.req.fire()) {
996d5ddbceSLemover        val req = io.req.bits
100b6982e83SLemover        state := s_addr_check
101cc5a5f22SLemover        level := Mux(req.l1Hit, 1.U, 0.U)
102b6982e83SLemover        af_level := Mux(req.l1Hit, 1.U, 0.U)
103cc5a5f22SLemover        ppn := Mux(req.l1Hit, io.req.bits.ppn, satp.ppn)
10445f497a4Shappy-lx        vpn := io.req.bits.req_info.vpn
1056d5ddbceSLemover        l1Hit := req.l1Hit
106b6982e83SLemover        accessFault := false.B
1076d5ddbceSLemover      }
1086d5ddbceSLemover    }
1096d5ddbceSLemover
110b6982e83SLemover    is (s_addr_check) {
111b6982e83SLemover      state := s_mem_req
112b6982e83SLemover    }
113b6982e83SLemover
1146d5ddbceSLemover    is (s_mem_req) {
1156d5ddbceSLemover      when (mem.req.fire()) {
1166d5ddbceSLemover        state := s_mem_resp
1176d5ddbceSLemover      }
118b6982e83SLemover      when (accessFault) {
119b6982e83SLemover        state := s_check_pte
120b6982e83SLemover      }
1216d5ddbceSLemover    }
1226d5ddbceSLemover
1236d5ddbceSLemover    is (s_mem_resp) {
1246d5ddbceSLemover      when(mem.resp.fire()) {
125cc5a5f22SLemover        state := s_check_pte
126b6982e83SLemover        af_level := af_level + 1.U
1276d5ddbceSLemover      }
1286d5ddbceSLemover    }
1296d5ddbceSLemover
130cc5a5f22SLemover    is (s_check_pte) {
1312b773508SZhangZifei      when (io.resp.valid) { // find pte already or accessFault (mentioned below)
1326d5ddbceSLemover        when (io.resp.fire()) {
1336d5ddbceSLemover          state := s_idle
1346d5ddbceSLemover        }
135b6982e83SLemover        finish := true.B
13692e3bfefSLemover      }.elsewhen(io.llptw.valid) { // the next level is pte, go to miss queue
13792e3bfefSLemover        when (io.llptw.fire()) {
138cc5a5f22SLemover          state := s_idle
139cc5a5f22SLemover        }
140b6982e83SLemover        finish := true.B
1412b773508SZhangZifei      } otherwise { // go to next level, access the memory, need pmp check first
1422b773508SZhangZifei        when (io.pmp.resp.ld) { // pmp check failed, raise access-fault
1432b773508SZhangZifei          // do nothing, RegNext the pmp check result and do it later (mentioned above)
1442b773508SZhangZifei        }.otherwise { // go to next level.
1452b773508SZhangZifei          assert(level === 0.U)
146b6982e83SLemover          level := levelNext
147b6982e83SLemover          state := s_mem_req
148cc5a5f22SLemover        }
149cc5a5f22SLemover      }
1506d5ddbceSLemover    }
1516d5ddbceSLemover  }
1526d5ddbceSLemover
1536d5ddbceSLemover  when (sfence.valid) {
1546d5ddbceSLemover    state := s_idle
155b6982e83SLemover    accessFault := false.B
1566d5ddbceSLemover  }
1576d5ddbceSLemover
158b6982e83SLemover  // memPte is valid when at s_check_pte. when mem.resp.fire, it's not ready.
159cc5a5f22SLemover  val is_pte = memPte.isLeaf() || memPte.isPf(level)
160cc5a5f22SLemover  val find_pte = is_pte
161cc5a5f22SLemover  val to_find_pte = level === 1.U && !is_pte
16245f497a4Shappy-lx  val source = RegEnable(io.req.bits.req_info.source, io.req.fire())
163b6982e83SLemover  io.resp.valid := state === s_check_pte && (find_pte || accessFault)
164cc5a5f22SLemover  io.resp.bits.source := source
16545f497a4Shappy-lx  io.resp.bits.resp.apply(pageFault && !accessFault, accessFault, Mux(accessFault, af_level, level), memPte, vpn, satp.asid)
166cc5a5f22SLemover
16792e3bfefSLemover  io.llptw.valid := state === s_check_pte && to_find_pte && !accessFault
16892e3bfefSLemover  io.llptw.bits.req_info.source := source
16992e3bfefSLemover  io.llptw.bits.req_info.vpn := vpn
17092e3bfefSLemover  io.llptw.bits.ppn := memPte.ppn
171cc5a5f22SLemover
172cc5a5f22SLemover  assert(level =/= 2.U || level =/= 3.U)
1736d5ddbceSLemover
1746d5ddbceSLemover  val l1addr = MakeAddr(satp.ppn, getVpnn(vpn, 2))
175cc5a5f22SLemover  val l2addr = MakeAddr(Mux(l1Hit, ppn, memPte.ppn), getVpnn(vpn, 1))
176b6982e83SLemover  val mem_addr = Mux(af_level === 0.U, l1addr, l2addr)
177b6982e83SLemover  io.pmp.req.valid := DontCare // samecycle, do not use valid
178b6982e83SLemover  io.pmp.req.bits.addr := mem_addr
179b6982e83SLemover  io.pmp.req.bits.size := 3.U // TODO: fix it
180b6982e83SLemover  io.pmp.req.bits.cmd := TlbCmd.read
181b6982e83SLemover
182b6982e83SLemover  mem.req.valid := state === s_mem_req && !io.mem.mask && !accessFault
183b6982e83SLemover  mem.req.bits.addr := mem_addr
184bc063562SLemover  mem.req.bits.id := FsmReqID.U(bMemID.W)
1856d5ddbceSLemover
18645f497a4Shappy-lx  io.refill.req_info.vpn := vpn
1876d5ddbceSLemover  io.refill.level := level
18845f497a4Shappy-lx  io.refill.req_info.source := source
1896d5ddbceSLemover
19092e3bfefSLemover  XSDebug(p"[ptw] state:${state} level:${level} notFound:${pageFault}\n")
1916d5ddbceSLemover
1926d5ddbceSLemover  // perf
1936d5ddbceSLemover  XSPerfAccumulate("fsm_count", io.req.fire())
1946d5ddbceSLemover  for (i <- 0 until PtwWidth) {
19545f497a4Shappy-lx    XSPerfAccumulate(s"fsm_count_source${i}", io.req.fire() && io.req.bits.req_info.source === i.U)
1966d5ddbceSLemover  }
1976d5ddbceSLemover  XSPerfAccumulate("fsm_busy", state =/= s_idle)
1986d5ddbceSLemover  XSPerfAccumulate("fsm_idle", state === s_idle)
1996d5ddbceSLemover  XSPerfAccumulate("resp_blocked", io.resp.valid && !io.resp.ready)
2006d5ddbceSLemover  XSPerfAccumulate("mem_count", mem.req.fire())
2016d5ddbceSLemover  XSPerfAccumulate("mem_cycle", BoolStopWatch(mem.req.fire, mem.resp.fire(), true))
2026d5ddbceSLemover  XSPerfAccumulate("mem_blocked", mem.req.valid && !mem.req.ready)
203cc5a5f22SLemover
2049bd9cdfaSLemover  TimeOutAssert(state =/= s_idle, timeOutThreshold, "page table walker time out")
205cd365d4cSrvcoresjw
206cd365d4cSrvcoresjw  val perfEvents = Seq(
207cd365d4cSrvcoresjw    ("fsm_count         ", io.req.fire()                                     ),
208cd365d4cSrvcoresjw    ("fsm_busy          ", state =/= s_idle                                  ),
209cd365d4cSrvcoresjw    ("fsm_idle          ", state === s_idle                                  ),
210cd365d4cSrvcoresjw    ("resp_blocked      ", io.resp.valid && !io.resp.ready                   ),
211cd365d4cSrvcoresjw    ("mem_count         ", mem.req.fire()                                    ),
212cd365d4cSrvcoresjw    ("mem_cycle         ", BoolStopWatch(mem.req.fire, mem.resp.fire(), true)),
213cd365d4cSrvcoresjw    ("mem_blocked       ", mem.req.valid && !mem.req.ready                   ),
214cd365d4cSrvcoresjw  )
2151ca0e4f3SYinan Xu  generatePerfEvent()
2166d5ddbceSLemover}
21792e3bfefSLemover
21892e3bfefSLemover/*========================= LLPTW ==============================*/
21992e3bfefSLemover
22092e3bfefSLemover/** LLPTW : Last Level Page Table Walker
22192e3bfefSLemover  * the page walker that only takes 4KB(last level) page walk.
22292e3bfefSLemover  **/
22392e3bfefSLemover
22492e3bfefSLemoverclass LLPTWInBundle(implicit p: Parameters) extends XSBundle with HasPtwConst {
22592e3bfefSLemover  val req_info = Output(new L2TlbInnerBundle())
22692e3bfefSLemover  val ppn = Output(UInt(PAddrBits.W))
22792e3bfefSLemover}
22892e3bfefSLemover
22992e3bfefSLemoverclass LLPTWIO(implicit p: Parameters) extends MMUIOBaseBundle with HasPtwConst {
23092e3bfefSLemover  val in = Flipped(DecoupledIO(new LLPTWInBundle()))
23192e3bfefSLemover  val out = DecoupledIO(new Bundle {
23292e3bfefSLemover    val req_info = Output(new L2TlbInnerBundle())
23392e3bfefSLemover    val id = Output(UInt(bMemID.W))
23492e3bfefSLemover    val af = Output(Bool())
23592e3bfefSLemover  })
23692e3bfefSLemover  val mem = new Bundle {
23792e3bfefSLemover    val req = DecoupledIO(new L2TlbMemReqBundle())
23892e3bfefSLemover    val resp = Flipped(Valid(new Bundle {
23992e3bfefSLemover      val id = Output(UInt(log2Up(l2tlbParams.llptwsize).W))
24092e3bfefSLemover    }))
24192e3bfefSLemover    val enq_ptr = Output(UInt(log2Ceil(l2tlbParams.llptwsize).W))
24292e3bfefSLemover    val buffer_it = Output(Vec(l2tlbParams.llptwsize, Bool()))
24392e3bfefSLemover    val refill = Output(new L2TlbInnerBundle())
24492e3bfefSLemover    val req_mask = Input(Vec(l2tlbParams.llptwsize, Bool()))
24592e3bfefSLemover  }
246*7797f035SbugGenerator  val cache = DecoupledIO(new L2TlbInnerBundle())
24792e3bfefSLemover  val pmp = new Bundle {
24892e3bfefSLemover    val req = Valid(new PMPReqBundle())
24992e3bfefSLemover    val resp = Flipped(new PMPRespBundle())
25092e3bfefSLemover  }
25192e3bfefSLemover}
25292e3bfefSLemover
25392e3bfefSLemoverclass LLPTWEntry(implicit p: Parameters) extends XSBundle with HasPtwConst {
25492e3bfefSLemover  val req_info = new L2TlbInnerBundle()
25592e3bfefSLemover  val ppn = UInt(ppnLen.W)
25692e3bfefSLemover  val wait_id = UInt(log2Up(l2tlbParams.llptwsize).W)
25792e3bfefSLemover  val af = Bool()
25892e3bfefSLemover}
25992e3bfefSLemover
26092e3bfefSLemover
26192e3bfefSLemover@chiselName
26292e3bfefSLemoverclass LLPTW(implicit p: Parameters) extends XSModule with HasPtwConst with HasPerfEvents {
26392e3bfefSLemover  val io = IO(new LLPTWIO())
26492e3bfefSLemover
265*7797f035SbugGenerator  val flush = io.sfence.valid || io.csr.satp.changed
26692e3bfefSLemover  val entries = Reg(Vec(l2tlbParams.llptwsize, new LLPTWEntry()))
267*7797f035SbugGenerator  val state_idle :: state_addr_check :: state_mem_req :: state_mem_waiting :: state_mem_out :: state_cache :: Nil = Enum(6)
26892e3bfefSLemover  val state = RegInit(VecInit(Seq.fill(l2tlbParams.llptwsize)(state_idle)))
269*7797f035SbugGenerator
27092e3bfefSLemover  val is_emptys = state.map(_ === state_idle)
27192e3bfefSLemover  val is_mems = state.map(_ === state_mem_req)
27292e3bfefSLemover  val is_waiting = state.map(_ === state_mem_waiting)
27392e3bfefSLemover  val is_having = state.map(_ === state_mem_out)
274*7797f035SbugGenerator  val is_cache = state.map(_ === state_cache)
27592e3bfefSLemover
27692e3bfefSLemover  val full = !ParallelOR(is_emptys).asBool()
27792e3bfefSLemover  val enq_ptr = ParallelPriorityEncoder(is_emptys)
27892e3bfefSLemover
279*7797f035SbugGenerator  val mem_ptr = ParallelPriorityEncoder(is_having) // TODO: optimize timing, bad: entries -> ptr -> entry
28092e3bfefSLemover  val mem_arb = Module(new RRArbiter(new LLPTWEntry(), l2tlbParams.llptwsize))
28192e3bfefSLemover  for (i <- 0 until l2tlbParams.llptwsize) {
28292e3bfefSLemover    mem_arb.io.in(i).bits := entries(i)
28392e3bfefSLemover    mem_arb.io.in(i).valid := is_mems(i) && !io.mem.req_mask(i)
28492e3bfefSLemover  }
28592e3bfefSLemover
286*7797f035SbugGenerator  val cache_ptr = ParallelMux(is_cache, (0 until l2tlbParams.llptwsize).map(_.U))
287*7797f035SbugGenerator
28892e3bfefSLemover  // duplicate req
28992e3bfefSLemover  // to_wait: wait for the last to access mem, set to mem_resp
29092e3bfefSLemover  // to_cache: the last is back just right now, set to mem_cache
29192e3bfefSLemover  val dup_vec = state.indices.map(i =>
29292e3bfefSLemover    dup(io.in.bits.req_info.vpn, entries(i).req_info.vpn)
29392e3bfefSLemover  )
29492e3bfefSLemover  val dup_req_fire = mem_arb.io.out.fire() && dup(io.in.bits.req_info.vpn, mem_arb.io.out.bits.req_info.vpn) // dup with the req fire entry
29592e3bfefSLemover  val dup_vec_wait = dup_vec.zip(is_waiting).map{case (d, w) => d && w} // dup with "mem_waiting" entres, sending mem req already
29692e3bfefSLemover  val dup_vec_having = dup_vec.zipWithIndex.map{case (d, i) => d && is_having(i)} // dup with the "mem_out" entry recv the data just now
29792e3bfefSLemover  val wait_id = Mux(dup_req_fire, mem_arb.io.chosen, ParallelMux(dup_vec_wait zip entries.map(_.wait_id)))
29892e3bfefSLemover  val dup_wait_resp = io.mem.resp.fire() && VecInit(dup_vec_wait)(io.mem.resp.bits.id) // dup with the entry that data coming next cycle
29992e3bfefSLemover  val to_wait = Cat(dup_vec_wait).orR || dup_req_fire
30092e3bfefSLemover  val to_mem_out = dup_wait_resp
301*7797f035SbugGenerator  val to_cache = Cat(dup_vec_having).orR
302*7797f035SbugGenerator  XSError(RegNext(dup_req_fire && Cat(dup_vec_wait).orR, init = false.B), "mem req but some entries already waiting, should not happed")
30392e3bfefSLemover
304*7797f035SbugGenerator  XSError(io.in.fire() && ((to_mem_out && to_cache) || (to_wait && to_cache)), "llptw enq, to cache conflict with to mem")
30592e3bfefSLemover  val mem_resp_hit = RegInit(VecInit(Seq.fill(l2tlbParams.llptwsize)(false.B)))
306*7797f035SbugGenerator  val enq_state_normal = Mux(to_mem_out, state_mem_out, // same to the blew, but the mem resp now
307*7797f035SbugGenerator    Mux(to_wait, state_mem_waiting,
308*7797f035SbugGenerator    Mux(to_cache, state_cache, state_addr_check)))
309*7797f035SbugGenerator  val enq_state = Mux(from_pre(io.in.bits.req_info.source) && enq_state_normal =/= state_addr_check, state_idle, enq_state_normal)
31092e3bfefSLemover  when (io.in.fire()) {
31192e3bfefSLemover    // if prefetch req does not need mem access, just give it up.
31292e3bfefSLemover    // so there will be at most 1 + FilterSize entries that needs re-access page cache
31392e3bfefSLemover    // so 2 + FilterSize is enough to avoid dead-lock
314*7797f035SbugGenerator    state(enq_ptr) := enq_state
31592e3bfefSLemover    entries(enq_ptr).req_info := io.in.bits.req_info
31692e3bfefSLemover    entries(enq_ptr).ppn := io.in.bits.ppn
31792e3bfefSLemover    entries(enq_ptr).wait_id := Mux(to_wait, wait_id, enq_ptr)
31892e3bfefSLemover    entries(enq_ptr).af := false.B
31992e3bfefSLemover    mem_resp_hit(enq_ptr) := to_mem_out
32092e3bfefSLemover  }
321*7797f035SbugGenerator
322*7797f035SbugGenerator  val enq_ptr_reg = RegNext(enq_ptr)
323*7797f035SbugGenerator  val need_addr_check = RegNext(enq_state === state_addr_check && io.in.fire() && !flush)
324*7797f035SbugGenerator  val last_enq_vpn = RegEnable(io.in.bits.req_info.vpn, io.in.fire())
325*7797f035SbugGenerator
326*7797f035SbugGenerator  io.pmp.req.valid := need_addr_check
327*7797f035SbugGenerator  io.pmp.req.bits.addr := RegEnable(MakeAddr(io.in.bits.ppn, getVpnn(io.in.bits.req_info.vpn, 0)), io.in.fire())
328*7797f035SbugGenerator  io.pmp.req.bits.cmd := TlbCmd.read
329*7797f035SbugGenerator  io.pmp.req.bits.size := 3.U // TODO: fix it
330*7797f035SbugGenerator  val pmp_resp_valid = io.pmp.req.valid // same cycle
331*7797f035SbugGenerator  when (pmp_resp_valid) {
332*7797f035SbugGenerator    // NOTE: when pmp resp but state is not addr check, then the entry is dup with other entry, the state was changed before
333*7797f035SbugGenerator    //       when dup with the req-ing entry, set to mem_waiting (above codes), and the ld must be false, so dontcare
334*7797f035SbugGenerator    val accessFault = io.pmp.resp.ld || io.pmp.resp.mmio
335*7797f035SbugGenerator    entries(enq_ptr_reg).af := accessFault
336*7797f035SbugGenerator    state(enq_ptr_reg) := Mux(accessFault, state_mem_out, state_mem_req)
337*7797f035SbugGenerator  }
338*7797f035SbugGenerator
33992e3bfefSLemover  when (mem_arb.io.out.fire()) {
34092e3bfefSLemover    for (i <- state.indices) {
34192e3bfefSLemover      when (state(i) =/= state_idle && dup(entries(i).req_info.vpn, mem_arb.io.out.bits.req_info.vpn)) {
34292e3bfefSLemover        // NOTE: "dup enq set state to mem_wait" -> "sending req set other dup entries to mem_wait"
34392e3bfefSLemover        state(i) := state_mem_waiting
34492e3bfefSLemover        entries(i).wait_id := mem_arb.io.chosen
34592e3bfefSLemover      }
34692e3bfefSLemover    }
34792e3bfefSLemover  }
34892e3bfefSLemover  when (io.mem.resp.fire()) {
34992e3bfefSLemover    state.indices.map{i =>
35092e3bfefSLemover      when (state(i) === state_mem_waiting && io.mem.resp.bits.id === entries(i).wait_id) {
35192e3bfefSLemover        state(i) := state_mem_out
35292e3bfefSLemover        mem_resp_hit(i) := true.B
35392e3bfefSLemover      }
35492e3bfefSLemover    }
35592e3bfefSLemover  }
35692e3bfefSLemover  when (io.out.fire()) {
35792e3bfefSLemover    assert(state(mem_ptr) === state_mem_out)
35892e3bfefSLemover    state(mem_ptr) := state_idle
35992e3bfefSLemover  }
36092e3bfefSLemover  mem_resp_hit.map(a => when (a) { a := false.B } )
36192e3bfefSLemover
362*7797f035SbugGenerator  when (io.cache.fire) {
363*7797f035SbugGenerator    state(cache_ptr) := state_idle
36492e3bfefSLemover  }
365*7797f035SbugGenerator  XSError(io.out.fire && io.cache.fire && (mem_ptr === cache_ptr), "mem resp and cache fire at the same time at same entry")
36692e3bfefSLemover
36792e3bfefSLemover  when (flush) {
36892e3bfefSLemover    state.map(_ := state_idle)
36992e3bfefSLemover  }
37092e3bfefSLemover
37192e3bfefSLemover  io.in.ready := !full
37292e3bfefSLemover
37392e3bfefSLemover  io.out.valid := ParallelOR(is_having).asBool()
37492e3bfefSLemover  io.out.bits.req_info := entries(mem_ptr).req_info
37592e3bfefSLemover  io.out.bits.id := mem_ptr
37692e3bfefSLemover  io.out.bits.af := entries(mem_ptr).af
37792e3bfefSLemover
37892e3bfefSLemover  io.mem.req.valid := mem_arb.io.out.valid && !flush
37992e3bfefSLemover  io.mem.req.bits.addr := MakeAddr(mem_arb.io.out.bits.ppn, getVpnn(mem_arb.io.out.bits.req_info.vpn, 0))
38092e3bfefSLemover  io.mem.req.bits.id := mem_arb.io.chosen
38192e3bfefSLemover  mem_arb.io.out.ready := io.mem.req.ready
38292e3bfefSLemover  io.mem.refill := entries(RegNext(io.mem.resp.bits.id(log2Up(l2tlbParams.llptwsize)-1, 0))).req_info
38392e3bfefSLemover  io.mem.buffer_it := mem_resp_hit
38492e3bfefSLemover  io.mem.enq_ptr := enq_ptr
38592e3bfefSLemover
386*7797f035SbugGenerator  io.cache.valid := Cat(is_cache).orR
387*7797f035SbugGenerator  io.cache.bits := ParallelMux(is_cache, entries.map(_.req_info))
388*7797f035SbugGenerator
38992e3bfefSLemover  XSPerfAccumulate("llptw_in_count", io.in.fire())
39092e3bfefSLemover  XSPerfAccumulate("llptw_in_block", io.in.valid && !io.in.ready)
39192e3bfefSLemover  for (i <- 0 until 7) {
39292e3bfefSLemover    XSPerfAccumulate(s"enq_state${i}", io.in.fire() && enq_state === i.U)
39392e3bfefSLemover  }
39492e3bfefSLemover  for (i <- 0 until (l2tlbParams.llptwsize + 1)) {
39592e3bfefSLemover    XSPerfAccumulate(s"util${i}", PopCount(is_emptys.map(!_)) === i.U)
39692e3bfefSLemover    XSPerfAccumulate(s"mem_util${i}", PopCount(is_mems) === i.U)
39792e3bfefSLemover    XSPerfAccumulate(s"waiting_util${i}", PopCount(is_waiting) === i.U)
39892e3bfefSLemover  }
39992e3bfefSLemover  XSPerfAccumulate("mem_count", io.mem.req.fire())
40092e3bfefSLemover  XSPerfAccumulate("mem_cycle", PopCount(is_waiting) =/= 0.U)
40192e3bfefSLemover  XSPerfAccumulate("blocked_in", io.in.valid && !io.in.ready)
40292e3bfefSLemover
40392e3bfefSLemover  for (i <- 0 until l2tlbParams.llptwsize) {
40492e3bfefSLemover    TimeOutAssert(state(i) =/= state_idle, timeOutThreshold, s"missqueue time out no out ${i}")
40592e3bfefSLemover  }
40692e3bfefSLemover
40792e3bfefSLemover  val perfEvents = Seq(
40892e3bfefSLemover    ("tlbllptw_incount           ", io.in.fire()               ),
40992e3bfefSLemover    ("tlbllptw_inblock           ", io.in.valid && !io.in.ready),
41092e3bfefSLemover    ("tlbllptw_memcount          ", io.mem.req.fire()          ),
41192e3bfefSLemover    ("tlbllptw_memcycle          ", PopCount(is_waiting)       ),
41292e3bfefSLemover  )
41392e3bfefSLemover  generatePerfEvent()
41492e3bfefSLemover}
415